We introduce LiveSecBench, a continuously updated safety benchmark specifically for Chinese-language LLM application scenarios. LiveSecBench constructs a high-quality and unique dataset through a pipeline that combines automated generation with human verification. By periodically releasing new versions to expand the dataset and update evaluation metrics, LiveSecBench provides a robust and up-to-date standard for AI safety. In this report, we introduce our second release v251215, which evaluates across five dimensions (Public Safety, Fairness & Bias, Privacy, Truthfulness, and Mental Health Safety.) We evaluate 57 representative LLMs using an ELO rating system, offering a leaderboard of the current state of Chinese LLM safety. The result is available at https://livesecbench.intokentech.cn/.
翻译:我们推出LiveSecBench,这是一个专门针对中文语言模型应用场景、持续更新的安全基准。LiveSecBench通过结合自动生成与人工验证的流程,构建了一个高质量且独特的数据集。通过定期发布新版本以扩展数据集并更新评估指标,LiveSecBench为人工智能安全提供了一个稳健且与时俱进的标准。本报告介绍了我们的第二个版本v251215,该版本在五个维度(公共安全、公平性与偏见、隐私、真实性及心理健康安全)进行评估。我们采用ELO评分系统对57个代表性语言模型进行了评估,提供了当前中文语言模型安全状况的排行榜。结果可在 https://livesecbench.intokentech.cn/ 查看。