From 1267833806d12490a10bb70fe7c124aaa2b010b9 Mon Sep 17 00:00:00 2001 From: Michael Luciuk Date: Mon, 7 Jul 2025 10:27:21 -0400 Subject: [PATCH] =?UTF-8?q?Add=20upload=20arrow=20=E2=AC=86=EF=B8=8F=20to?= =?UTF-8?q?=20each=20upload?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- .riahub/workflows/workflow.yaml | 44 ++++++++++++--------------------- 1 file changed, 16 insertions(+), 28 deletions(-) diff --git a/.riahub/workflows/workflow.yaml b/.riahub/workflows/workflow.yaml index d422ada..ae49e06 100644 --- a/.riahub/workflows/workflow.yaml +++ b/.riahub/workflows/workflow.yaml @@ -1,4 +1,4 @@ -name: RIA Hub Workflow Demo +name: Modulation Recognition Demo on: push: @@ -11,9 +11,6 @@ on: jobs: ria-demo: runs-on: ubuntu-latest-2080 - env: - RIAGIT_USERNAME: ${{ secrets.USERNAME }} - RIAGIT_TOKEN: ${{ secrets.TOKEN }} steps: - name: Print GPU information run: | @@ -24,7 +21,7 @@ jobs: echo "⚠️ No NVIDIA GPU found" fi - - name: Checkout code + - name: Checkout project code uses: actions/checkout@v4 with: lfs: true @@ -42,13 +39,10 @@ jobs: utils \ -r requirements.txt - - - name: 1. Generate Recordings run: | mkdir -p data/recordings - PYTHONPATH=. python scripts/dataset_building/data_gen.py --output-dir data/recordings - echo "recordings produced successfully" + PYTHONPATH=. python scripts/dataset_manager/data_gen.py --output-dir data/recordings - name: ⬆️ Upload recordings uses: actions/upload-artifact@v3 @@ -59,11 +53,10 @@ jobs: - name: 2. Build HDF5 Dataset run: | mkdir -p data/dataset - PYTHONPATH=. python scripts/dataset_building/produce_dataset.py - echo "datasets produced successfully" + PYTHONPATH=. python scripts/dataset_manager/produce_dataset.py shell: bash - - name: 📤 Upload Dataset + - name: ⬆️ Upload Dataset uses: actions/upload-artifact@v3 with: name: dataset @@ -75,34 +68,30 @@ jobs: PYTORCH_NO_NNPACK: 1 run: | mkdir -p checkpoint_files - PYTHONPATH=. python scripts/training/train.py 2>/dev/null - echo "training model" + PYTHONPATH=. python scripts/model_builder/train.py 2>/dev/null - name: 4. Plot Model env: NO_NNPACK: 1 PYTORCH_NO_NNPACK: 1 run: | - PYTHONPATH=. python scripts/training/plot_data.py 2>/dev/null - + PYTHONPATH=. python scripts/model_builder/plot_data.py 2>/dev/null - - name: Upload Checkpoints + - name: ⬆️ Upload Checkpoints uses: actions/upload-artifact@v3 with: name: checkpoints path: checkpoint_files/* - - - name: 5. Convert to ONNX file + - name: 5. Export model to ONNX graph env: NO_NNPACK: 1 PYTORCH_NO_NNPACK: 1 run: | mkdir -p onnx_files - MKL_DISABLE_FAST_MM=1 PYTHONPATH=. python scripts/onnx/convert_to_onnx.py 2>/dev/null - echo "building inference app" + MKL_DISABLE_FAST_MM=1 PYTHONPATH=. python scripts/application_packager/convert_to_onnx.py 2>/dev/null - - name: Upload ONNX file + - name: ⬆️ Upload ONNX file uses: actions/upload-artifact@v3 with: name: onnx-file @@ -110,21 +99,20 @@ jobs: - name: 6. Profile ONNX model run: | - PYTHONPATH=. python scripts/onnx/profile_onnx.py + PYTHONPATH=. python scripts/application_packager/profile_onnx.py - - name: Upload JSON profiling data + - name: ⬆️ Upload JSON trace uses: actions/upload-artifact@v3 with: name: profile-data path: '**/onnxruntime_profile_*.json' - - name: 7. Convert to ORT file + - name: 7. Convert ONNX graph to an ORT file run: | - PYTHONPATH=. python scripts/ort/convert_to_ort.py + PYTHONPATH=. python scripts/application_packager/convert_to_ort.py - - name: Upload ORT file + - name: ⬆️ Upload ORT file uses: actions/upload-artifact@v3 with: name: ort-file path: ort_files/inference_recognition_model.ort -