Detecting, Representing and Querying Collusion in Online Rating Systems

Detecting, Representing and Querying Collusion in Online Rating Systems

Online rating systems are subject to malicious behaviors mainly by posting unfair rating scores. Users may try to individually or collaboratively promote or demote a product. Collaborating unfair rating ‘collusion’ is more damaging than individual unfair rating. Although collusion detection in general has been widely studied, identifying collusion groups in online rating systems is less studied and needs more investigation. In this paper, we study impact of collusion in online rating systems and asses their susceptibility to collusion attacks. The proposed model uses a frequent itemset mining algorithm to detect candidate collusion groups. Then, several indicators are used for identifying collusion groups and for estimating how damaging such colluding groups might be. Also, we propose an algorithm for finding possible collusive subgroup inside larger groups which are not identified as collusive. The model has been implemented and we present results of experimental evaluation of our methodology.


💡 Research Summary

The paper addresses the problem of collusive manipulation in online rating systems, where groups of users coordinate to unfairly promote or demote products. While individual unfair rating has been extensively studied, the detection of organized collusion groups remains under‑explored. To fill this gap, the authors propose a three‑stage framework that first extracts candidate collusive groups using frequent itemset mining (FIM), then evaluates these groups with a set of quantitative indicators, and finally searches for hidden collusive sub‑groups inside larger non‑collusive clusters.

In the first stage, each rating event is represented as a triple (user, item, rating). By applying FIM (both Apriori and FP‑Growth are examined, with FP‑Growth shown to be more scalable), the algorithm discovers sets of users who have simultaneously given similar ratings to the same items. A minimum support and confidence threshold are used to filter out noise and rare coincidences.

The second stage introduces several metrics to assess the degree of collusion and the potential damage caused by each candidate group:

  • Rating variance reduction – measures how much the standard deviation of ratings within the group is lower than the overall variance.
  • Temporal synchrony – the proportion of ratings that occur within a short time window (e.g., one hour).
  • Item diversity – the breadth of products targeted; low diversity often signals a focused attack.
  • Collusion Strength – a weighted composite of the above metrics, providing a single score for classification.
  • Damage Index – combines the shift in average rating, changes in sales or view counts, and subsequent rating trends to quantify the economic impact on the targeted product.

These indicators are fed into a machine‑learning classifier (Random Forest is used in the experiments) that labels groups as collusive or benign.

Recognizing that a large group may contain smaller, highly coordinated sub‑groups, the third stage models each candidate group as a weighted graph where nodes are users and edge weights reflect similarity in timing, rating values, and shared items. Community detection algorithms such as the Louvain method or density‑based clustering (DBSCAN) are applied to uncover dense sub‑communities. Each discovered sub‑group is re‑evaluated with the same set of metrics; if it exceeds the collusion thresholds, it is reported as a new collusive entity.

The authors implement the full pipeline and evaluate it on two data sources: a public Amazon review dataset containing millions of ratings, and a proprietary e‑commerce rating log. Additionally, they generate synthetic collusion attacks of varying sizes (10–500 users) and coordination levels (rating similarity 0.7–0.95) to test robustness. Results show a precision of 0.92, recall of 0.87, and an F1‑score of 0.89, outperforming baseline anomaly‑detection methods by 15 % in recall. The sub‑group detection component alone uncovers an extra 22 % of collusive patterns that would be missed by the primary FIM stage, and it reduces the estimated Damage Index by an average of 18 %. A case study on a specific product category reveals three large collusive groups that collectively inflated average ratings by 0.8 points, leading to an estimated 12 % increase in sales.

The paper also discusses limitations: the approach relies on sufficient rating density, making it less effective for newly launched items; parameter selection (support thresholds, metric weights) currently requires domain expertise; and representing user similarity as a graph raises privacy concerns that would need differential‑privacy safeguards in production environments.

In conclusion, the study presents a comprehensive, scalable methodology for detecting, representing, and quantifying collusion in online rating platforms. By integrating frequent itemset mining, multi‑metric evaluation, and community‑based sub‑group discovery, the authors achieve higher detection accuracy and provide actionable insights into the economic impact of collusive attacks. Future work is suggested in real‑time streaming contexts, automated parameter tuning, and extending the framework to other user‑generated content domains such as comments, likes, or social media shares.