Please use this identifier to cite or link to this item: https://ah.lib.nccu.edu.tw/handle/140.119/84114
題名: Inferring Potential Users in Mobile Social Networks
作者: Hsu, T.-H.;Chen, C.-C.;Chiang, M.-F.;Hsu, K.-W.;Peng, W.-C.
徐國偉
貢獻者: 資科系
日期: Oct-2014
上傳時間: 11-Apr-2016
摘要: In mobile social networks, users can communicate with each other over different telecom operators. Thus, for telecom operators, how to attract new customers is a significant issue. The work of churn prediction is to determine whether a customer would leave soon. Differing from churn prediction, our work is to find those users who are likely to join target services from the competitors in the near future, where these users are called potential users. To infer potential users, we propose a framework including feature extraction, feature selection, and classifier learning to solve the problem. First, we construct a heterogeneous information network from the call detail records of users. Then, we extract the explicit features from potential users` interaction behavior in the heterogeneous information network. Moreover, because users are influenced by their community, we extract community-based implicit features of potential users. After feature extraction, we explore the Information Gain to select the effective features. We use the effective explicit and implicit features to learn potential user classifiers, and use the classifiers to determine the potential users. Finally, we conduct experiments on real datasets. The results of our experiments show that the features extracted by our proposed method can improve the accuracy of inferring potential users.
關聯: International Conference on Data Science and Advanced Analytics (DSAA), Shanghai, China, October 30-November 1, 2014, 347-353
資料類型: conference
DOI: http://dx.doi.org/10.1109/DSAA.2014.7058095
Appears in Collections:會議論文

Files in This Item:
File Description SizeFormat
index.html220 BHTML2View/Open
Show full item record

Google ScholarTM

Check

Altmetric

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.