[关键词]
[摘要]
数据异常尚没有统一的定义,其含义是指可能破坏数据库一致性状态的特定数据操作模式.已知的数据异常有脏写、脏读、不可重复读、幻读、丢失更新、读偏序和写偏序等.为了提高并发控制算法的效率,数据异常也被用于定义隔离级别,采用较弱的隔离级别以提高事务处理系统的效率.体系化地研究了数据异常以及对应的隔离级别,发现了22种未被其他文献报告过的新的数据异常,并对全部数据异常进行分类.基于数据异常的分类,提出了新的且不同粒度的隔离级别体系,揭示基于数据异常定义隔离级别的规律,使得对于数据异常和隔离级别等相关概念的认知可以更加简明.
[Key word]
[Abstract]
There is no unified definition of data anomalies, which refers to the specific data operation mode that may destroy the consistency of the database. Known data anomalies include Dirty Write, Dirty Read, Non-repeatable Read, Phantom, Read Skew, Write Skew, etc. In order to improve the efficiency of concurrency control algorithms, data anomalies are also used to define the isolation levels, because the weak isolation level can improve the efficiency of transaction processing systems. This work systematically studies the data anomalies and the corresponding isolation levels. Twenty-two new data anomalies are reported that have not been reported by other researches, and all data anomalies are classified miraculously. Based on the classification of data anomalies, two new isolation levels with different granularity are proposed, which reveals the rule of defining isolation levels based on data anomalies and makes the cognition of data anomalies and isolation levels more concise.
[中图分类号]
[基金项目]
国家重点研发计划(2017YFB1001803);国家自然科学基金(61872008)