RIA_Example/README.md
2026-04-22 09:21:42 -04:00

13 KiB

RIA Hub — Example Files

This repository contains example input files, configurations, and expected outputs for every tool on RIA Hub. If you are new to the platform, start here. Download any file and follow the walkthrough for the tool you want to try.


What is RIA Hub?

RIA Hub is a collaborative platform for RF and machine learning workflows. It combines a Git-based repository system with a suite of specialized tools that cover the full pipeline from raw IQ recordings to live inference deployments:

Stage Tools
Collect Library — browse, organize, and share RF recordings and models
Curate Dataset Manager — slice, qualify, augment, and inspect radio datasets
Train Model Builder — train, optimize, and compress PyTorch models
Deploy Application Packager — compose and build inference applications
Run Screens — deploy live RF inference pipelines on real hardware

Repository Structure

RIA_Example/
│
├── recordings/
│   └── example_iq_recording.h5          # Raw IQ capture (input to Curator)
│
├── datasets/
│   ├── example_radio_dataset.h5          # Curated radio dataset (Curator output / Model Trainer input)
│   └── example_synthetic_dataset.h5      # Synthetically generated dataset (Generator output)
│
├── models/
│   ├── example_model.ckpt                  # PyTorch Module (Model Trainer input / output)
│   └── example_model.onnx               # Exported ONNX model (Screens / Application Packager input)
│
├── applications/
│   └── example_application.json         # Application Composer output, can be built in RIA Screens (Application Packager)
│

│
├── .ria/
│   ├── train.yaml                        # Example Model Trainer workflow (committed to .ria/)
│   ├── example_application.yaml          # Example Application Composer Build workflow (committed to .ria/)

│
└── curator-configs/
    └── example_curator_config.json       # Example curation configuration for the Curator tool

Tool Walkthroughs

Library

The Library is a cross-repository browser for all RF and ML assets on the platform. It automatically discovers files pushed to any repository you have access to.

To explore the example recording:

  1. Import recordings/example_iq_recording.h5 into any repository via New Repository → Upload Files or by pushing via Git LFS.
  2. Navigate to Library in the top navigation bar.
  3. Select the Recordings tab. Your file will appear with metadata and a spectrogram thumbnail.
  4. Click the file to open the detail view — you can inspect signal properties, view the spectrogram, and copy the file to another repository.

Supported asset types in the Library:

Type Extension Description
Recording .h5 / .hdf5 Raw IQ capture files
Radio Dataset .h5 / .hdf5 Labelled, curated training datasets
PyTorch Module .py PyTorch model definitions with a nn.Module class
PyTorch State Dict .pt / .pth Model weights / state dictionaries
PyTorch Checkpoint .ckpt Training checkpoints with weights, optimizer state, and metadata
ONNX Graph .onnx Portable inference models

Dataset Manager — Curator

The Curator takes raw IQ recordings and produces a labelled, ready-to-train HDF5 dataset. It applies a configurable DSP pipeline: slicing, quality filtering, and optional augmentation.

Example files: recordings/example_iq_recording.h5, curator-configs/example_curator_config.json
Expected output: datasets/example_radio_dataset.h5

Steps:

  1. Upload example_iq_recording.h5 to a repository (the Curator reads from the Library).
  2. Go to Dataset Manager → Curator.
  3. Select your recording from the Library panel on the left.
  4. Configure the pipeline using the settings below, or load example_curator_config.json as a reference:
    • Data type: IQ
    • Slicer: simple — slice length 1024
    • Qualifier: rms — minimum threshold 0.01 (filters out silent/noise-only slices)
    • Augmentation: basic policy — 2 augmented copies per slice
  5. Set a dataset name, description, and radio task label.
  6. Click Curate. A progress bar tracks the Celery task.
  7. When complete, commit the output dataset to your repository.

Slicer options:

Slicer Best for
simple Fixed-length slices, good starting point
random Randomized slice positions
overlap Overlapping slices for smaller datasets

Qualifier options:

Qualifier Filters on
rms Root mean square amplitude
snr Signal-to-noise ratio estimate
energy Total signal energy
bandwidth Occupied bandwidth

Dataset Manager — Inspector

The Inspector runs diagnostic analysis on an existing dataset — class balance, per-class statistics, anomaly detection, and dataset comparisons.

Example file: datasets/example_radio_dataset.h5

Steps:

  1. Go to Dataset Manager → Inspector.
  2. Select example_radio_dataset.h5 from the file picker.
  3. Choose an analysis type:
    • Balance — see how many samples exist per class label
    • Per-Class Stats — per-class mean, std, and distribution
    • Anomaly Detection — flag outlier samples
    • Compare — select a second dataset to diff against

Dataset Manager — Generator

The Generator creates synthetic labelled datasets from a parameter sweep without requiring any hardware or recordings.

Expected output: datasets/example_synthetic_dataset.h5

Steps:

  1. Go to Dataset Manager → Generator.
  2. Configure a modulation sweep:
    • Sampling strategy: grid
    • Parameters: SNR from -5 to 20 dB in steps of 5; modulation types: [BPSK, QPSK, 8PSK, 16QAM]
    • Signal: length 1024, sample rate 1e6
    • Channel model: awgn
    • Output backend: pytorch
  3. Click Generate. The task runs in the background.
  4. Download the resulting .h5 file when complete.

Model Builder — Model Trainer

The Model Trainer builds a training workflow YAML and commits it to your repository. A Gitea Actions runner then executes the training job.

Example files: datasets/example_radio_dataset.h5, models/example_model.ckpt (optional pre-trained start)
Expected output: .riahub/workflows/train.yaml in your repository, plus a trained example_model.ckpt artifact

Steps:

  1. Go to Model Builder → Model Trainer.
  2. In Repository, select the repository where you want to store the workflow and output artifacts.
  3. In Model, choose an architecture (e.g. ResNet1D) or use example_model.ckpt as a starting checkpoint.
  4. In Dataset, select example_radio_dataset.h5 from the Library.
  5. Configure training:
    • Optimizer: Adam, learning rate 1e-3
    • Epochs: 20
    • Batch size: 64
    • Criterion: CrossEntropyLoss
  6. Enable ONNX Export in the Evaluation section to automatically export the trained model.
  7. Click Commit Workflow. A train.yaml is committed to .riahub/workflows/ and a CI run starts.
  8. Monitor the run in Actions within your repository.

The committed workflow file matches workflows/train.yaml in this repository.


Model Builder — Hyperparameter Optimization

HPO runs a sweep over a configurable search space, training multiple model variants and ranking them by a target metric.

Example file: datasets/example_radio_dataset.h5
Expected output: .riahub/workflows/hpo.yaml

Steps:

  1. Go to Model Builder → HPO.
  2. Configure the same model and dataset as in Model Trainer.
  3. In the Search Space panel, define ranges to sweep:
    • Learning rate: 1e-4 to 1e-2 (log scale)
    • Optimizer: [Adam, SGD]
    • Batch size: [32, 64, 128]
  4. Set Trials: 12 and Target metric: val_accuracy.
  5. Click Commit Workflow. See workflows/hpo.yaml for the expected output format.

Model Builder — Model Compression

Compression applies pruning and/or quantization to reduce model size for edge deployment. The output is an ONNX file.

Example files: models/example_model.ckpt, datasets/example_radio_dataset.h5
Expected output: models/example_model.onnx

Steps:

  1. Go to Model Builder → Compression.
  2. Select example_model.ckpt as the source model and example_radio_dataset.h5 as the calibration dataset.
  3. Configure the compression pipeline (pruning ratio, quantization bits).
  4. Click Commit Workflow. The Actions job exports the compressed model to ONNX automatically.
  5. The resulting .onnx file is committed back to your repository.

Application Packager — Application Composer

The Application Composer is a visual node-graph editor for wiring together C++ operator blocks into an inference application. The output is an application JSON file.

Example file: applications/example_application.json

Steps:

  1. Go to Application Packager → Application Composer.
  2. Browse the Operators panel on the left. Drag an operator onto the canvas.
  3. Wire operator ports together by dragging from an output port to an input port.
  4. Configure each operator's parameters in the sidebar.
  5. Click Commit Application to save example_application.json to your repository.
  6. Click Build to trigger a build workflow on a registered runner.

The application JSON format is documented in schemas/application/ria_application.schema.json. See applications/example_application.json for a minimal working example.

Target profiles:

Profile Use when
native-x86 Standard x86 Linux deployment
native-arm64 ARM edge devices
nvidia-x86 GPU-accelerated inference on x86
RIA Screens Web based runtime environment to monitor app

Screens

Screens deploys a packaged RF inference application to a live pipeline. You build an app from Application Composer, configure a data source (live SDR, file playback, or synthetic), and start the pipeline. Results stream back to the browser in real time.

App package format

A Screens app package is a .tar.gz containing:

  • manifest.json — describes the app (models, GUI layout, data source, preprocessor)
  • ONNX model file(s) at the path(s) listed in manifest.models[].path

Data source types:

Type Description
synthetic Built-in AWGN tone generator — no hardware required
recording Play back a .h5 IQ recording from the Library
sdr Live data from a connected SDR device
agent Live data from a remote SDR via an edge agent node
numpy_raw Play back a .npy raw IQ file

ONNX model requirements

The Zone Fingerprinting model contract:

  • Input: iq_features — shape [1, 128], dtype float32
  • Output: scores — shape [1, 5], dtype float32 (softmax probabilities)
  • Opset: >= 13

When building your own Screens app, export your model to ONNX with matching input/output names and shapes, then reference them in manifest.json.


RIA Projects

Projects group your datasets, models, training runs, and deployed applications into a single tracked entity. The project dashboard shows a three-stage pipeline view: Data Management → Model Building → Deployment.

Steps:

  1. Go to Projects → New Project.
  2. Name your project and create it.
  3. Link assets from the Library using the Link Asset button on each pipeline stage.
  4. As you run Curator, Model Trainer, and Screens jobs, link the outputs to track progress through the pipeline.

File Format Reference

HDF5 Radio Dataset (.h5)

Curated and generated datasets share a common HDF5 layout:

dataset.h5
├── data/           # IQ samples, shape [N, slice_length, 2] (float32)
├── labels/         # Integer class labels, shape [N]
├── metadata/       # Recording metadata carried through from source
└── attrs           # Dataset-level attributes: name, version, radio_task, backend

Application JSON (application.json)

{
  "app_name": "my_inference_app",
  "backend": "native",
  "target_profile": "native-x86",
  "ops": [
    {
      "name": "source",
      "class_name": "UDPSourceOp",
      "type": "source",
      "inputs": [],
      "outputs": [{ "name": "output" }],
      "specs": [
        { "name": "port", "value": "5000", "arg_type": "int" }
      ]
    }
  ],
  "flows": [
    {
      "upstream": "source",
      "downstream": "inference",
      "port_pairs": { "output": "input" }
    }
  ]
}

Getting Help

  • Full platform documentation is available in the Docs section of RIA Hub.
  • Open an issue in this repository if an example file is missing, broken, or out of date.
  • For tool-specific questions, use the in-app help panels (the ? icon on each tool page).