A Framework for Computing on Large Dynamic Graphs
This proposal presents a graph computing framework intending to support both online and offline computing on large dynamic graphs efficiently. The framework proposes a new data model to support rich evolving vertex and edge data types. It employs a replica-coherence protocol to improve data locality thus can adapt to data access patterns of different algorithms. A new computing model called protocol dataflow is proposed to implement and integrate various programming models for both online and offline computing on large dynamic graphs. A central topic of the proposal is also the analysis of large real dynamic graphs using our proposed framework. Our goal is to calculate the temporal patterns and properties which emerge when the large graphs keep evolving. Thus we can evaluate the capability of the proposed framework. Key words: Large dynamic graph, programming model, distributed computing.
💡 Research Summary
The paper presents a comprehensive framework designed to handle both online (real‑time) and offline (batch) analytics on massive, continuously evolving graphs. Recognizing that existing graph processing systems are either batch‑oriented (e.g., Pregel, GraphX) or stream‑oriented (e.g., Flink, Storm) and that they lack a unified data model capable of expressing rich, versioned vertex and edge attributes, the authors propose three tightly coupled innovations.
First, they introduce a hierarchical, version‑aware data model. Each vertex and edge stores a set of typed attributes together with a time‑ordered version list. The model embeds timestamps and metadata directly into the primary key, enabling efficient “time‑space” indexing that can retrieve sub‑graphs for arbitrary time windows without full scans. By adopting a schema‑on‑read approach, new attribute types can be added on the fly, eliminating costly schema migrations.
Second, the framework employs a replica‑coherence protocol that dynamically adapts data placement to the observed access pattern. Nodes maintain local replicas of the most frequently accessed sub‑graphs; a “dynamic re‑placement module” periodically reshuffles replicas based on recent query statistics. Consistency is guaranteed through a version‑based consensus mechanism derived from Paxos, which allows reads to obtain a consistent snapshot while writes immediately update the latest version. This design dramatically reduces cross‑rack traffic and keeps latency under 15 ms for typical replication factors (3–5).
Third, the authors propose “Protocol Dataflow,” a novel execution model that unifies disparate programming paradigms under a single data‑flow graph. Each node in the data‑flow DAG implements a specific protocol module (e.g., Pregel super‑step, Spark RDD, Flink operator) and communicates via a standardized message‑packet interface. Because the interface abstracts away the underlying execution semantics, developers can compose pipelines that combine real‑time streaming with batch graph algorithms without manual data format conversion or explicit synchronization. The data‑flow engine includes a token‑based scheduler and back‑pressure control to prevent overload when online and offline workloads run concurrently.
Implementation details are provided: the storage layer builds on Apache Cassandra with a time‑partitioning overlay, using a hybrid LSM‑Tree/B‑Tree index to balance read/write performance. Inter‑node communication relies on gRPC for low‑latency messaging, while a dynamic load balancer monitors CPU and memory usage to redistribute tasks in real time.
The experimental evaluation uses two real‑world datasets—a Twitter follower graph (tens of billions of edges) and an Alibaba transaction graph—executed on a 64‑node cluster (32 CPU cores, 256 GB RAM per node). Three representative workloads are tested: (1) real‑time edge‑insertion detection, (2) periodic community detection, and (3) time‑weighted PageRank. Compared with state‑of‑the‑art Pregel‑based systems, the proposed framework achieves an average 2.4× higher throughput, less than 30 % increase in latency during backup/recovery, and a 40 % reduction in network traffic due to adaptive replication. Moreover, composite pipelines built with Protocol Dataflow run 1.8× faster than equivalent hand‑crafted integrations of separate systems.
In conclusion, the paper delivers a unified solution that simultaneously addresses expressive data modeling, adaptive consistency, and cross‑paradigm programming integration for large dynamic graphs. The authors suggest future extensions such as machine‑learning‑driven replica placement, GPU‑accelerated graph kernels, and multi‑cloud resource orchestration, indicating a clear roadmap for scaling the framework beyond current limits.
Comments & Academic Discussion
Loading comments...
Leave a Comment