论文标题
在网络钓鱼电子邮件检测中对联邦学习的评估
Evaluation of Federated Learning in Phishing Email Detection
论文作者
论文摘要
使用人工智能(AI)来检测网络钓鱼电子邮件主要取决于大规模的集中数据集,这使其始于无数的隐私,信任和法律问题。此外,考虑到商业上敏感信息泄漏的风险,组织不愿分享电子邮件。因此,获得足够的电子邮件来有效培训全球AI模型并不常见。因此,保存隐私的分布式和协作机器学习,尤其是联合学习(FL),这是一个逃避者。在医疗保健领域已经很普遍,在多组织合作的背景下,基于FL的网络钓鱼检测的有效性和功效仍然存在问题。据我们所知,这里的工作是第一个研究FL在电子邮件反捕捞中使用的工作。本文建立在深度神经网络模型的基础上,尤其是用于网络钓鱼电子邮件检测的RNN和BERT。它分析了各种环境(包括平衡和不对称数据分布)的范围内的学习绩效。我们的结果证实了网络钓鱼电子邮件检测中FL的可比性能统计数据,即平衡数据集的集中学习和较低的组织计数。此外,在增加组织计数时,我们会观察到绩效的变化。对于固定的总电子邮件数据集,基于全球RNN的模型在将组织计数从2增加到10时遭受1.8%的精度下降。相比之下,伯特准确度从2个组织到5个组织时上升了0.6%。但是,如果我们允许通过在FL框架中引入新组织来增加整体电子邮件数据集,则通过实现更快的收敛速度来提高组织级别的绩效。此外,如果电子邮件数据集发行非常不对称,则FL由于非常不稳定的输出而遭受了整体全球模型性能。
The use of Artificial Intelligence (AI) to detect phishing emails is primarily dependent on large-scale centralized datasets, which opens it up to a myriad of privacy, trust, and legal issues. Moreover, organizations are loathed to share emails, given the risk of leakage of commercially sensitive information. So, it is uncommon to obtain sufficient emails to train a global AI model efficiently. Accordingly, privacy-preserving distributed and collaborative machine learning, particularly Federated Learning (FL), is a desideratum. Already prevalent in the healthcare sector, questions remain regarding the effectiveness and efficacy of FL-based phishing detection within the context of multi-organization collaborations. To the best of our knowledge, the work herein is the first to investigate the use of FL in email anti-phishing. This paper builds upon a deep neural network model, particularly RNN and BERT for phishing email detection. It analyzes the FL-entangled learning performance under various settings, including balanced and asymmetrical data distribution. Our results corroborate comparable performance statistics of FL in phishing email detection to centralized learning for balanced datasets, and low organization counts. Moreover, we observe a variation in performance when increasing organizational counts. For a fixed total email dataset, the global RNN based model suffers by a 1.8% accuracy drop when increasing organizational counts from 2 to 10. In contrast, BERT accuracy rises by 0.6% when going from 2 to 5 organizations. However, if we allow increasing the overall email dataset with the introduction of new organizations in the FL framework, the organizational level performance is improved by achieving a faster convergence speed. Besides, FL suffers in its overall global model performance due to highly unstable outputs if the email dataset distribution is highly asymmetric.