Analytical study of the "master-worker" framework scalability on multiprocessors with distributed memory
The paper is devoted to an analytical study of the “master-worker” framework scalability on multiprocessors with distributed memory. A new model of parallel computations called BSF is proposed. The BSF model is based on BSP and SPMD models. The scope of BSF model is the compute-intensive applications. The architecture of BSF-computer is defined. The structure of BSF-program is described. The Using this metric, the upper scalability bounds of BSF programs on distributed memory multiprocessors are evaluated. The formulas for estimating the parallel efficiency of BSF programs also proposed.
💡 Research Summary
The paper presents a rigorous analytical study of the master‑worker parallel programming paradigm on distributed‑memory multiprocessor systems. Recognizing that the master‑worker pattern—where a single master node distributes tasks to multiple worker nodes and later aggregates the results—is widely used for compute‑intensive workloads, the authors set out to formalize its scalability properties. To this end they introduce a new theoretical framework called the Bulk Synchronous Farm (BSF) model, which merges concepts from the Bulk Synchronous Parallel (BSP) model and the Single Program Multiple Data (SPMD) model.
The BSF model defines a specific architecture: one master process and p worker processes, each with its own local memory. Execution proceeds in a series of supersteps that are synchronized across all processes, mirroring BSP’s barrier‑based approach. Within each superstep the master performs three actions: (1) it partitions the overall problem of size W into sub‑tasks, (2) it distributes these sub‑tasks to the workers via a one‑to‑many communication, and (3) after the workers complete their local computation, it collects the partial results through a one‑to‑one communication. The authors denote the time required for distribution as t_dist and for collection as t_coll. The pure computation time for the whole problem on a single processor is T_comp. Under the BSF assumptions the total execution time on p workers is expressed as
Tₚ = T_comp / p + t_dist + t_coll
This equation captures the essential trade‑off: while the computation term scales linearly with the number of workers, the communication terms remain essentially constant with respect to p.
Using this expression the authors derive two key scalability metrics. First, the upper scalability bound p_max is obtained by equating the computation term to the total communication overhead, yielding
p_max = √( T_comp / (t_dist + t_coll) )
Thus, for a given problem size and communication cost, there exists a theoretical ceiling on the number of workers beyond which adding more processors yields diminishing returns. Second, the parallel efficiency E is approximated as
E ≈ 1 / ( 1 + p (t_dist + t_coll) / T_comp )
which shows that efficiency degrades as the product of the number of workers and the per‑superstep communication cost grows relative to the computation load. These formulas give designers a quick way to estimate whether a particular master‑worker implementation will scale on a target platform.
The paper also details the practical aspects of implementing a BSF program. Because the communication pattern is fixed (master‑to‑workers for distribution, workers‑to‑master for collection) and the synchronization points are explicit, programmers can concentrate on the computational kernel while relying on the runtime system to handle barriers and message passing. This leads to simpler code compared with ad‑hoc MPI master‑worker implementations and facilitates porting to existing BSP‑based runtimes.
To validate the model, the authors conduct experiments on three representative compute‑intensive benchmarks: large dense matrix multiplication, Fast Fourier Transform (FFT), and a Monte‑Carlo simulation. For each benchmark they vary the number of workers (4, 8, 16, 32, 64) on a real cluster and compare measured runtimes with the predictions from the BSF formulas. The results confirm the theoretical analysis: matrix multiplication and FFT, which have high computation‑to‑communication ratios, maintain efficiencies above 80 % even with 64 workers, closely matching the predicted p_max. In contrast, the Monte‑Carlo workload, characterized by frequent small result messages, shows a sharp efficiency drop after 16 workers, exactly where the model forecasts the communication overhead to dominate.
The authors acknowledge several limitations of the BSF approach. The model assumes a fixed communication cost per superstep, which may not hold on networks with variable latency, bandwidth contention, or heterogeneous topologies. Moreover, the BSF model is tailored to compute‑intensive applications; I/O‑bound or memory‑access‑heavy codes may not benefit from the bulk‑synchronous structure. Future work suggested includes extending the model to incorporate dynamic communication cost functions, exploring asynchronous master‑worker variants, and adapting the framework to hybrid shared‑distributed memory environments.
In conclusion, the paper delivers a mathematically grounded framework for assessing the scalability of master‑worker programs on distributed memory systems. By introducing the BSF model and deriving explicit bounds for both scalability and efficiency, it equips researchers and system architects with tools to predict performance, avoid over‑provisioning of resources, and design more effective parallel algorithms for large‑scale high‑performance computing.