Skip to content

Ensemble ML for RF Signal Classification: A Reproducible Performance Study


Smarter RF Signal Classification: How Ensemble ML Boosts Accuracy and Reliability

By Benjamin J. Gilbert – Spectrcyde RF Quantum SCYTHE, College of the Mainland

In noisy RF environments, where modulation signals blur under interference and fading, a single classifier often isn’t enough. Enter ensemble machine learning—a strategy that combines multiple models into a smarter whole.

Our latest reproducible study benchmarks lightweight ensembles that blend deep learning and traditional ML for automatic modulation recognition (AMR). The results show clear wins for confidence-weighted voting, feature fusion, and careful calibration.


Why Ensemble Learning for RF?

RF classification is notoriously brittle:

  • Noise & interference degrade accuracy fast.
  • Different modulations overlap in spectral/temporal space.
  • One-size-fits-all models often overfit specific conditions.

Ensembles address these pain points by diversifying learners. Think of it like a jury: different members see different clues, and their combined decision is more robust.


What We Tested

We built reproducible ensembles with:

  • Voting Schemes: Majority vs. confidence-weighted voting.
  • Feature Fusion: Combining spectral & temporal features.
  • Classical Models: Random Forests & SVMs alongside CNNs.

Performance was measured across seven modulation classes (AM, FM, SSB, CW, PSK, FSK, Noise) under SNRs from −5 to +15 dB.

Key metrics:

  • Accuracy & Macro-F1 – how well models classify.
  • Latency – runtime per sample.
  • ECE (Expected Calibration Error) – how trustworthy the probabilities are.

Results in Brief

  • Weighted voting wins: Outperformed majority voting across all SNR conditions.
  • Feature fusion helps: Modest accuracy gains, but with added compute cost.
  • Classical ML shines in the noise: Random Forest-like integration improved results at low SNR.
  • Calibration is essential: Post-hoc calibration reduced ECE significantly, turning raw confidence into reliable risk scores.

📊 At 10 dB SNR, the top configuration (weighted voting + fusion + traditional ML) scored 83.5% accuracy, F1 = 0.835, with low ECE = 0.05—all in ~3.2 ms per sample.


Why It Matters

For real-world RF applications—telecoms, defense, IoT—the findings point to three practical takeaways:

  1. Don’t rely on one model: Ensembles provide resilience across variable SNRs.
  2. Calibrate everything: A confident but wrong classifier is worse than an uncertain but honest one.
  3. Balance cost vs. performance: Feature fusion improves performance, but you need to weigh latency budgets.

Reproducibility by Design

Every step of the pipeline is open and deterministic:

  • Fixed seeds for synthetic IQ data.
  • JSON-tracked metrics for comparison.
  • Automated TeX → PDF reporting.
  • Pure Python, cross-platform builds.

This ensures results can be verified, extended, or directly integrated into new RF classification research.


Conclusion

Ensemble ML methods are not just an academic curiosity—they’re a practical path to more robust RF signal recognition. With confidence-weighted voting, feature fusion, and calibration, we can build classifiers that don’t just perform better—they know when to trust themselves.

📡 Bottom line: Smarter ensembles mean smarter spectrum sensing—paving the way for more reliable wireless systems under real-world noise.


Leave a Reply

Your email address will not be published. Required fields are marked *