|
| 1 | +--- |
| 2 | +title: 'FairSynergy: Fair Resource Allocation for Fleet Intelligence' |
| 3 | +description: GLOBECOM 2025 paper |
| 4 | +categories: blog |
| 5 | +--- |
| 6 | + |
| 7 | +*By Oguzhan Baser* |
| 8 | + |
| 9 | +# FairSynergy: Fair Resource Allocation for Fleet Intelligence |
| 10 | + |
| 11 | +2025 IEEE Global Communications Conference (GLOBECOM 2025) |
| 12 | + |
| 13 | +**TLDR:** Give or substitute each extra unit of compute or memory to the agent that benefits most to lift fleet accuracy by up to +25% in inference and +11% in training, with even bigger gains as fleets grow! |
| 14 | + |
| 15 | +[arXiv](https://arxiv.org/abs/2509.03353) | |
| 16 | +[Code](https://github.com/UTAustin-SwarmLab/Fair-Synergy) |
| 17 | + |
| 18 | +## Motivation |
| 19 | +Intelligent fleets involving machine learning (ML) deployed drones, autonomous cars, or phones rarely look uniform. Some fleet agents have much better on-board compute or memory, while others face harder inputs such as hard-to-parse images and complex prompts. While cloud assists fleets by allocating its limited resources, it is not really trivial how to allocate for optimal collective performance across the heterogeneous agents. If we split cloud resources evenly, we waste the budget on saturated parts of the accuracy-vs-resource curve while starving steep ones. We ask: |
| 20 | + |
| 21 | +- **Q1 (Fairness):** Where does the next unit of resource raise accuracy the most? |
| 22 | +- **Q2 (Multivariate Utility):** How can we substitute multiple types of ML resources available? |
| 23 | + |
| 24 | + |
| 25 | +<figure style="text-align: center;"> |
| 26 | + <img src="{{site.baseurl}}/images/post/FS_sysarch.png" alt="FairSynergy System Plot" height="auto" style="margin: auto; display: block;"> |
| 27 | + <figcaption>FairSynergy Overview</figcaption> |
| 28 | + <p></p> |
| 29 | +</figure> |
| 30 | + |
| 31 | + |
| 32 | +## Contributions |
| 33 | +We introduce FairSynergy, a novel framework to allocate cloud resources fairly across the intelligent heterogeneous agents. |
| 34 | +- **Diminishing Marginal Returns:** We empirically show that learning and inference curves across vision and language models consistently have the same concave pattern: accuracy improves with more resources (e.g., model size, data, compute) but with diminishing returns. |
| 35 | +- **Fair Allocation in ML:** With a concave objective, Network Utility Maximization (NUM) is the natural fit. It mathematically implies the condition turns into an intuitive policy: allocate so each agent’s next unit of resource yields the same marginal accuracy gain. |
| 36 | +- **Cobb-Douglas:** Like capital and labor in production, compute/model capacity and labeling/data curation both drive accuracy, each with diminishing returns and substitution. We form a multivariate utility captures this to co-allocate the resources. (Figs. 5–6) |
| 37 | + |
| 38 | +<figure style="text-align: center;"> |
| 39 | + <img src="{{site.baseurl}}/images/post/FS_concavetraining.png" alt="Results Boxplot" height="auto" style="margin: auto; display: block;"> |
| 40 | + <figcaption>The Law of Diminishing Marginal Returns: Training</figcaption> |
| 41 | + <p></p> |
| 42 | +</figure> |
| 43 | + |
| 44 | +<figure style="text-align: center;"> |
| 45 | + <img src="{{site.baseurl}}/images/post/FS_result_concaveinference.png" alt="Results Boxplot" height="auto" style="margin: auto; display: block;"> |
| 46 | + <figcaption>The Law of Diminishing Marginal Returns: Inference</figcaption> |
| 47 | + <p></p> |
| 48 | +</figure> |
| 49 | + |
| 50 | +## FairSynergy Framework |
| 51 | + |
| 52 | +- **Inference Setting (RTI)- Univariate Case (Compute):** At short intervals, the framwork estimates each agent’s next-unit accuracy gain from extra cloud compute. Give the next unit to the highest gain, repeat until gains are roughly equalized—then reshuffle as conditions change. This hits the fairness/efficiency sweet spot without heavy tuning. |
| 53 | + |
| 54 | + |
| 55 | +- **Learning Setting (DL) - Bivariate Case (Compute + Labeling Effort):** The framework uses the same “next-unit” idea with a quick two-step loop: hold labels fixed and split compute by the one resource rule; then hold compute fixed and split labeling time by the same rule. A few rounds settle to a stable co-allocation so compute-hungry agents get cycles and data-hungry agents get labels. |
| 56 | + |
| 57 | +- **Handling Heterogeneity:** Harder tasks show larger early gains, so the allocator leans into them first and naturally rebalances as gains even out. The result is proportional fairness and fleet-level accuracy that scales with more agents and changing workloads—no math knobs to tune. |
| 58 | + |
| 59 | +<figure style="text-align: center;"> |
| 60 | + <img src="{{site.baseurl}}/images/post/FS_result_concaveinference.png" alt="Results Boxplot" height="auto" style="margin: auto; display: block;"> |
| 61 | + <figcaption>**Extending Multivariate ML Utility**: Cobb-Douglas Production Function For a Given Capital and Labor</figcaption> |
| 62 | + <p></p> |
| 63 | +</figure> |
| 64 | + |
| 65 | +## Results |
| 66 | + |
| 67 | + |
| 68 | +We compare our method to common baselines and standard fair allocation methods: |
| 69 | + |
| 70 | +- **Fair-Synergy (Ours)** allocates compute (and labels) to equalize next-unit accuracy gains per agent using fitted concave utilities. |
| 71 | + |
| 72 | +- **Random Allocation** splits the available compute (and labels) at random among agents. |
| 73 | + |
| 74 | +- **Uniform Allocation** gives every agent the same share of compute (and labels), ignoring local differences. |
| 75 | + |
| 76 | +- **Classical NUM** optimizes a uniform log utility (identical response curves), so allocation follows equalized marginal gains without task-specific reweighting and agent heterogeneity. |
| 77 | + |
| 78 | +- **Dominant Resource Fairness (DRF)** equalizes each agent’s dominant resource share, targeting max–min fairness across resources. |
| 79 | + |
| 80 | +- **Leximin** prioritizes the worst-off first, maximizing the minimum utility, then the next, and so on. It is a stricted form of fairness compared to other methods. |
| 81 | + |
| 82 | +<figure style="text-align: center;"> |
| 83 | + <img src="{{site.baseurl}}/images/post/FS_result_boxplot.png" alt="Results Boxplot" height="auto" style="margin: auto; display: block;"> |
| 84 | + <figcaption>How well does Fair-Synergy perform compared to the benchmarks?</figcaption> |
| 85 | + <p></p> |
| 86 | +</figure> |
| 87 | + |
| 88 | +<figure style="text-align: center;"> |
| 89 | + <img src="{{site.baseurl}}/images/post/FS_result_lineplot.png" alt="Results Lineplots" height="auto" style="margin: auto; display: block;"> |
| 90 | + <figcaption>How does Fair-Synergy scale with increasing number of agents?</figcaption> |
| 91 | + <p></p> |
| 92 | +</figure> |
| 93 | + |
| 94 | +### Key findings: |
| 95 | +- **Fairness-Accuracy Trade-off:** We observe for stricter fairness conditions, while the worst case accuracy increases the aggregate accuracy decreases. Hovewer, we achieve the best trade-off by obtaining both the best performance and necessary fairness conditions. |
| 96 | +- **Scalability:** As fleets grow, our method converts heterogeneity into higher accuracy unlike (+6% RTI, +30% DL) and Uniform/DRF/Leximin (+13% RTI, +51% DL) at 100 agents. |
| 97 | + |
| 98 | + |
| 99 | +## Impact |
| 100 | +Fair-Synergy treats fairness as physics, not philosophy: if accuracy is concave, the right thing is to spend cloud resources where marginal gains are steepest and to do so jointly over multiple substitutable resources. A fair allocation is the most efficient allocation because concavity makes “equalize marginal gains” optimal. |
0 commit comments