13 KiB
RIA Hub — Example Files
This repository contains example input files, configurations, and expected outputs for every tool on RIA Hub. If you are new to the platform, start here. Download any file and follow the walkthrough for the tool you want to try.
What is RIA Hub?
RIA Hub is a collaborative platform for RF and machine learning workflows. It combines a Git-based repository system with a suite of specialized tools that cover the full pipeline from raw IQ recordings to live inference deployments:
| Stage | Tools |
|---|---|
| Collect | Library — browse, organize, and share RF recordings and models |
| Curate | Dataset Manager — slice, qualify, augment, and inspect radio datasets |
| Train | Model Builder — train, optimize, and compress PyTorch models |
| Deploy | Application Packager — compose and build inference applications |
| Run | Screens — deploy live RF inference pipelines on real hardware |
Repository Structure
RIA_Example/
│
├── recordings/
│ └── example_iq_recording.h5 # Raw IQ capture (input to Curator)
│
├── datasets/
│ ├── example_radio_dataset.h5 # Curated radio dataset (Curator output / Model Trainer input)
│ └── example_synthetic_dataset.h5 # Synthetically generated dataset (Generator output)
│
├── models/
│ ├── example_model.ckpt # PyTorch Module (Model Trainer input / output)
│ └── example_model.onnx # Exported ONNX model (Screens / Application Packager input)
│
├── applications/
│ └── example_application.json # Application Composer output, can be built in RIA Screens (Application Packager)
│
│
├── .ria/
│ ├── train.yaml # Example Model Trainer workflow (committed to .ria/)
│ ├── example_application.yaml # Example Application Composer Build workflow (committed to .ria/)
│
└── curator-configs/
└── example_curator_config.json # Example curation configuration for the Curator tool
Tool Walkthroughs
Library
The Library is a cross-repository browser for all RF and ML assets on the platform. It automatically discovers files pushed to any repository you have access to.
To explore the example recording:
- Import
recordings/example_iq_recording.h5into any repository via New Repository → Upload Files or by pushing via Git LFS. - Navigate to Library in the top navigation bar.
- Select the Recordings tab. Your file will appear with metadata and a spectrogram thumbnail.
- Click the file to open the detail view — you can inspect signal properties, view the spectrogram, and copy the file to another repository.
Supported asset types in the Library:
| Type | Extension | Description |
|---|---|---|
| Recording | .h5 / .hdf5 |
Raw IQ capture files |
| Radio Dataset | .h5 / .hdf5 |
Labelled, curated training datasets |
| PyTorch Module | .py |
PyTorch model definitions with a nn.Module class |
| PyTorch State Dict | .pt / .pth |
Model weights / state dictionaries |
| PyTorch Checkpoint | .ckpt |
Training checkpoints with weights, optimizer state, and metadata |
| ONNX Graph | .onnx |
Portable inference models |
Dataset Manager — Curator
The Curator takes raw IQ recordings and produces a labelled, ready-to-train HDF5 dataset. It applies a configurable DSP pipeline: slicing, quality filtering, and optional augmentation.
Example files: recordings/example_iq_recording.h5, curator-configs/example_curator_config.json
Expected output: datasets/example_radio_dataset.h5
Steps:
- Upload
example_iq_recording.h5to a repository (the Curator reads from the Library). - Go to Dataset Manager → Curator.
- Select your recording from the Library panel on the left.
- Configure the pipeline using the settings below, or load
example_curator_config.jsonas a reference:- Data type:
IQ - Slicer:
simple— slice length1024 - Qualifier:
rms— minimum threshold0.01(filters out silent/noise-only slices) - Augmentation:
basicpolicy —2augmented copies per slice
- Data type:
- Set a dataset name, description, and radio task label.
- Click Curate. A progress bar tracks the Celery task.
- When complete, commit the output dataset to your repository.
Slicer options:
| Slicer | Best for |
|---|---|
simple |
Fixed-length slices, good starting point |
random |
Randomized slice positions |
overlap |
Overlapping slices for smaller datasets |
Qualifier options:
| Qualifier | Filters on |
|---|---|
rms |
Root mean square amplitude |
snr |
Signal-to-noise ratio estimate |
energy |
Total signal energy |
bandwidth |
Occupied bandwidth |
Dataset Manager — Inspector
The Inspector runs diagnostic analysis on an existing dataset — class balance, per-class statistics, anomaly detection, and dataset comparisons.
Example file: datasets/example_radio_dataset.h5
Steps:
- Go to Dataset Manager → Inspector.
- Select
example_radio_dataset.h5from the file picker. - Choose an analysis type:
- Balance — see how many samples exist per class label
- Per-Class Stats — per-class mean, std, and distribution
- Anomaly Detection — flag outlier samples
- Compare — select a second dataset to diff against
Dataset Manager — Generator
The Generator creates synthetic labelled datasets from a parameter sweep without requiring any hardware or recordings.
Expected output: datasets/example_synthetic_dataset.h5
Steps:
- Go to Dataset Manager → Generator.
- Configure a modulation sweep:
- Sampling strategy:
grid - Parameters: SNR from
-5to20dB in steps of5; modulation types:[BPSK, QPSK, 8PSK, 16QAM] - Signal: length
1024, sample rate1e6 - Channel model:
awgn - Output backend:
pytorch
- Sampling strategy:
- Click Generate. The task runs in the background.
- Download the resulting
.h5file when complete.
Model Builder — Model Trainer
The Model Trainer builds a training workflow YAML and commits it to your repository. A Gitea Actions runner then executes the training job.
Example files: datasets/example_radio_dataset.h5, models/example_model.ckpt (optional pre-trained start)
Expected output: .riahub/workflows/train.yaml in your repository, plus a trained example_model.ckpt artifact
Steps:
- Go to Model Builder → Model Trainer.
- In Repository, select the repository where you want to store the workflow and output artifacts.
- In Model, choose an architecture (e.g.
ResNet1D) or useexample_model.ckptas a starting checkpoint. - In Dataset, select
example_radio_dataset.h5from the Library. - Configure training:
- Optimizer:
Adam, learning rate1e-3 - Epochs:
20 - Batch size:
64 - Criterion:
CrossEntropyLoss
- Optimizer:
- Enable ONNX Export in the Evaluation section to automatically export the trained model.
- Click Commit Workflow. A
train.yamlis committed to.riahub/workflows/and a CI run starts. - Monitor the run in Actions within your repository.
The committed workflow file matches workflows/train.yaml in this repository.
Model Builder — Hyperparameter Optimization
HPO runs a sweep over a configurable search space, training multiple model variants and ranking them by a target metric.
Example file: datasets/example_radio_dataset.h5
Expected output: .riahub/workflows/hpo.yaml
Steps:
- Go to Model Builder → HPO.
- Configure the same model and dataset as in Model Trainer.
- In the Search Space panel, define ranges to sweep:
- Learning rate:
1e-4to1e-2(log scale) - Optimizer:
[Adam, SGD] - Batch size:
[32, 64, 128]
- Learning rate:
- Set Trials:
12and Target metric:val_accuracy. - Click Commit Workflow. See
workflows/hpo.yamlfor the expected output format.
Model Builder — Model Compression
Compression applies pruning and/or quantization to reduce model size for edge deployment. The output is an ONNX file.
Example files: models/example_model.ckpt, datasets/example_radio_dataset.h5
Expected output: models/example_model.onnx
Steps:
- Go to Model Builder → Compression.
- Select
example_model.ckptas the source model andexample_radio_dataset.h5as the calibration dataset. - Configure the compression pipeline (pruning ratio, quantization bits).
- Click Commit Workflow. The Actions job exports the compressed model to ONNX automatically.
- The resulting
.onnxfile is committed back to your repository.
Application Packager — Application Composer
The Application Composer is a visual node-graph editor for wiring together C++ operator blocks into an inference application. The output is an application JSON file.
Example file: applications/example_application.json
Steps:
- Go to Application Packager → Application Composer.
- Browse the Operators panel on the left. Drag an operator onto the canvas.
- Wire operator ports together by dragging from an output port to an input port.
- Configure each operator's parameters in the sidebar.
- Click Commit Application to save
example_application.jsonto your repository. - Click Build to trigger a build workflow on a registered runner.
The application JSON format is documented in schemas/application/ria_application.schema.json. See applications/example_application.json for a minimal working example.
Target profiles:
| Profile | Use when |
|---|---|
native-x86 |
Standard x86 Linux deployment |
native-arm64 |
ARM edge devices |
nvidia-x86 |
GPU-accelerated inference on x86 |
RIA Screens |
Web based runtime environment to monitor app |
Screens
Screens deploys a packaged RF inference application to a live pipeline. You build an app from Application Composer, configure a data source (live SDR, file playback, or synthetic), and start the pipeline. Results stream back to the browser in real time.
App package format
A Screens app package is a .tar.gz containing:
manifest.json— describes the app (models, GUI layout, data source, preprocessor)- ONNX model file(s) at the path(s) listed in
manifest.models[].path
Data source types:
| Type | Description |
|---|---|
synthetic |
Built-in AWGN tone generator — no hardware required |
recording |
Play back a .h5 IQ recording from the Library |
sdr |
Live data from a connected SDR device |
agent |
Live data from a remote SDR via an edge agent node |
numpy_raw |
Play back a .npy raw IQ file |
ONNX model requirements
The Zone Fingerprinting model contract:
- Input:
iq_features— shape[1, 128], dtypefloat32 - Output:
scores— shape[1, 5], dtypefloat32(softmax probabilities) - Opset: >= 13
When building your own Screens app, export your model to ONNX with matching input/output names and shapes, then reference them in manifest.json.
RIA Projects
Projects group your datasets, models, training runs, and deployed applications into a single tracked entity. The project dashboard shows a three-stage pipeline view: Data Management → Model Building → Deployment.
Steps:
- Go to Projects → New Project.
- Name your project and create it.
- Link assets from the Library using the Link Asset button on each pipeline stage.
- As you run Curator, Model Trainer, and Screens jobs, link the outputs to track progress through the pipeline.
File Format Reference
HDF5 Radio Dataset (.h5)
Curated and generated datasets share a common HDF5 layout:
dataset.h5
├── data/ # IQ samples, shape [N, slice_length, 2] (float32)
├── labels/ # Integer class labels, shape [N]
├── metadata/ # Recording metadata carried through from source
└── attrs # Dataset-level attributes: name, version, radio_task, backend
Application JSON (application.json)
{
"app_name": "my_inference_app",
"backend": "native",
"target_profile": "native-x86",
"ops": [
{
"name": "source",
"class_name": "UDPSourceOp",
"type": "source",
"inputs": [],
"outputs": [{ "name": "output" }],
"specs": [
{ "name": "port", "value": "5000", "arg_type": "int" }
]
}
],
"flows": [
{
"upstream": "source",
"downstream": "inference",
"port_pairs": { "output": "input" }
}
]
}
Getting Help
- Full platform documentation is available in the Docs section of RIA Hub.
- Open an issue in this repository if an example file is missing, broken, or out of date.
- For tool-specific questions, use the in-app help panels (the
?icon on each tool page).