-
Notifications
You must be signed in to change notification settings - Fork 0
random forest
GPU-accelerated Random Forest ensemble trained via cuML, producing one model per forecast horizon (6 total). Excels at short-term predictions (Lead 1–2).
| Parameter | Value |
|---|---|
| Library | cuML (NVIDIA RAPIDS) |
| Trees per model | 50 |
| Models | 6 (one per Lead 1–6) |
| Input | Flat feature vector from Panel Construction |
| Training | GPU-accelerated |
Unlike the sequence models (LSTM, Transformer), each RF model is trained independently on a specific lead target, receiving the full 18-feature lag vector as a flat input.
- Best short-term F1: 0.97 at Lead 1
- Highest recall across short horizons — catches nearly all genuine surges
- Robust: No hyperparameter sensitivity, handles missing features gracefully
- Fast training: GPU acceleration via cuML
- RMSE degrades at Lead 4–6 (long-horizon patterns are harder for tree ensembles)
- No sequence modeling: Treats lag features independently rather than as a temporal sequence
- 6 separate models: No parameter sharing across horizons
The Horizon-Aware Ensemble gives RF the highest weight at short horizons:
| Lead | RF Weight |
|---|---|
| 1 | 60% |
| 2 | 50% |
| 3 | 30% |
| 4 | 20% |
| 5 | 10% |
| 6 | 5% |
| Lead | F1 | Precision | Recall | RMSE |
|---|---|---|---|---|
| 1 | 0.97 | 0.96 | 0.97 | Best at lead 1 |
| 6 | ~0.82 | ~0.78 | ~0.87 | Degrades |
See Model Performance for full comparison.
src/models/trained_models/
├── rf_lead_1.joblib
├── rf_lead_2.joblib
├── rf_lead_3.joblib
├── rf_lead_4.joblib
├── rf_lead_5.joblib
└── rf_lead_6.joblib
- Horizon-Aware Ensemble — How RF is weighted in production
- LSTM — Mid-horizon alternative
- Transformer — Long-horizon alternative
- Training Pipeline — Training process
- GPU Acceleration — cuML infrastructure
- Project Overview — Goals, research questions, methodology, and team
- Glossary — Key terms used throughout this wiki
Raw inputs that feed the prediction system.
| Page | Description |
|---|---|
| Visa Data | US Department of State visa issuance statistics (108 monthly PDFs) |
| Encounter Data | CBP Southwest border encounter statistics (FY2019–2026) |
| Google News | 170K+ news articles across 15 countries × 8 topics |
| Google Trends | Monthly search-interest time series (15 countries × 8 keywords) |
| Exchange Rates | IMF Real Effective Exchange Rate for 6 countries |
The end-to-end flow from raw data to production forecasts.
| Page | Description |
|---|---|
| Data Collection | Ingestion layer: async scraping, bounded concurrency, retry logic |
| Data Processing | PDF parsing, JSON→Parquet, encounter merging |
| NLP Enrichment | Embedding → Clustering → Labeling → Sentiment |
| Panel Construction | Feature engineering: 18 lag features, 6 lead targets |
| Training Pipeline | Out-of-time train/test split, 4 architectures |
| Inference Pipeline | Horizon-aware ensemble, production prediction flow |
Machine learning architectures and their roles in the ensemble.
| Page | Description |
|---|---|
| Random Forest | cuML GPU Random Forest — best at short horizons (Lead 1–2) |
| LSTM | MigrationLSTM — country-aware with SurgeJointLoss |
| Transformer | MigrationTransformer — best at long horizons (Lead 5–6) |
| Horizon-Aware Ensemble | Dynamic weighting: RF→short, Transformer→long |
| SurgeJointLoss | Dual-objective loss: Huber + BCE for crisis detection |
| Jina v5 Embeddings | TensorRT INT8 news article embeddings (768-dim) |
| Flan-T5 Summarization | TensorRT INT8 cluster labeling engine |
Statistical techniques driving the lead-lag and surge analysis.
| Page | Description |
|---|---|
| Lead-Lag Analysis | Pearson correlation at 0–6 month offsets |
| Surge Detection | Quantile-based and σ-threshold spike identification |
| Sentiment Analysis | Rule-based lexicon scoring for migration-relevant news |
| Event Clustering | HDBSCAN GPU clustering + LED label generation |
| Cross-Correlation Analysis | CCF analysis, VAR benchmarking, ADF stationarity tests |
| Multiple Comparison Correction | Benjamini-Hochberg FDR for 58 significant signals |
What the system discovered about migration predictability.
| Page | Description |
|---|---|
| Event-Visa Findings | News events as leading indicators (r=0.617 at 3-month lag) |
| Exchange Rate Findings | Exchange rate signals (DR r=0.498 at 2-month lag) |
| Model Performance | Ensemble results: F1=0.96 at Lead 1, F1=0.86 at Lead 6 |
Reference documentation for every src/ subpackage and key files.
| Page | Description |
|---|---|
| Main Entry Point |
src/main.py CLI: bootstrap, collect-live, sync-data |
| Collection Module |
src/collection/* — visa, encounter, news, trends, HF sync |
| Processing Module |
src/processing/* — parse, merge, build_panel, summarize |
| Analysis Module |
src/analysis/* — events, exchange_rate, trends_analysis, plots |
| Models Module |
src/models/* — surge_model, train_and_evaluate, inference |
| News Scraper | Deep dive: batch decoding, checkpoint recovery, throttling |
| PDF Parser | Deep dive: PyMuPDF table extraction, VISA_MAP normalization |
| TensorRT Engines | Deep dive: Jina-v5, Flan-T5, LED TensorRT engines |
| Build Panel Detail | Deep dive: lag/lead construction, forward-fill strategies |
| HF Sync | Deep dive: bidirectional Hugging Face Hub sync |
Compute, reproducibility, and operational details.
| Page | Description |
|---|---|
| GPU Acceleration | TensorRT INT8, cuML, CUDA streams, NVML profiling |
| Reproducibility | HF bootstrap, run.sh pipeline, dependency checking |