Register for High-Frequency Market Predictor

Predicts short-term price movement from tick data

Finance
Benchmark Requirements
Proposal ID
12
Model Name
High-Frequency Market Predictor
Token Ticker
HIFRE
Benchmark Metric
Sharpe Ratio
Target / Baseline Value
2.1
Dataset / Test Data Reference
kaggle
Token Reward
1,000 HIFRE per verified DeltaOne
Performance target

Reach 2.1 on Sharpe Ratio where higher is better.

Risk-adjusted return metric

How to Submit Your Model
1

Register Your Model

Use the proposal-owned values below to attach your model registration to the existing proposal through the SDK.

Install SDK

pip install "hokusai-ml-platform[ml]"

Configure API Key

export HOKUSAI_API_KEY="your-hokusai-api-key-here"

Register your model

hokusai model register \
  --token-id HIFRE \
  --model-path ./models/final_model.pkl \
  --metric sharpe_ratio \
  --baseline 0.65

Register High-Frequency Market Predictor with token HIFRE

import os
import mlflow
from hokusai.core import ModelRegistry

# Set up MLflow tracking URI
mlflow.set_tracking_uri("https://registry.hokus.ai/api/mlflow")

# IMPORTANT: Use your Hokusai API key, NOT an MLflow token
# The Hokusai API key authenticates both the registry and MLflow
os.environ["MLFLOW_TRACKING_TOKEN"] = os.getenv("HOKUSAI_API_KEY")

# Initialize registry
registry = ModelRegistry()

# Register your model
with mlflow.start_run() as run:
    # Log your model (replace with your actual model)
    mlflow.sklearn.log_model(
        your_trained_model,
        "model",
        registered_model_name="High-Frequency Market Predictor"
    )

    # Register with Hokusai
    model_uri = f"runs:/{run.info.run_id}/model"
    registered_model = registry.register_tokenized_model(
        model_uri=model_uri,
        model_name="High-Frequency Market Predictor",
        token_id="HIFRE",
        metric_name="sharpe_ratio",
        baseline_value=0.92,
        additional_tags={"proposal_id": "12", "dataset_ref": "kaggle", "author": "your-name", "version": "1.0"}
    )

    print(f"✅ Model registered successfully: {registered_model.name}")
2

Trigger Evaluation

Once your model is registered, trigger an evaluation run against the benchmark dataset. The system will automatically measure your model's performance against the target criteria.

Evaluation process:

  • Your model runs on the specified evaluation dataset
  • Performance is measured using Sharpe Ratio
  • Results are verified and recorded on-chain
  • Token rewards are calculated based on performance improvement
3

Claim Your Rewards

If your model meets or exceeds the benchmark target, you'll earn token rewards. Tokens are automatically minted and can be claimed from your dashboard.

What to Expect

Evaluation Timeline

Evaluation runs typically complete within 5-30 minutes, depending on model complexity and dataset size. You'll receive notifications when evaluation completes.

Performance Verification

All evaluation results are verified and recorded on-chain to ensure transparency and prevent manipulation. Your model's performance will be publicly visible.

Token Distribution

Tokens are minted automatically when your model achieves performance improvements. The amount is calculated based on the delta between your model's performance and the baseline, multiplied by the tokens-per-delta-one rate.

Support & Documentation

Need help? Visit our model submission guide for detailed instructions, code examples, and troubleshooting tips.