拉尿有泡沫是什么原因| 7岁属什么| 向日葵的花语是什么| 劳碌命是什么意思| 生冷辛辣指的是什么| 刺猬爱吃什么| 为什么越累越胖| 培土什么意思| 梦见自己买衣服是什么意思| 阴道瘙痒吃什么药| 外阴痒是什么原因| 莲子心和什么搭配泡水喝最好| 三点水一个分读什么| 北海龙王叫什么| 杀马特是什么| 艾滋病吃什么药| 款式是什么意思| 5月23是什么星座| 发烧吃什么水果好| 梦到头发白了是什么意思| 肩胛骨疼挂什么科| 坐月子能吃什么蔬菜| 玄学是什么| 六九是什么意思| 百合有什么功效| 聿五行属什么| 经期吃什么| 夸张是什么意思| 痰涎壅盛是什么意思| 酵母菌属于什么菌| 安排是什么意思| 转氨酶异常是什么意思| 天蝎座与什么星座最配| 吃虫草有什么好处| 指压板有什么功效| 见利忘义是什么意思| 回族信奉什么教| 女生补肾吃什么| 面部痉挛是什么原因引起的| 焚香是什么意思| 区局长是什么级别| 母亲节送母亲什么礼物| 海肠是什么动物| 1901年属什么生肖| 男人左眼皮跳是什么预兆| 手上掉皮什么原因| cm什么单位| 珩字五行属什么| 公安局跟派出所有什么区别| 同型半胱氨酸是什么意思| 体型最大的恐龙是什么| 坚壁清野什么意思| 囚徒是什么意思| 禁忌症是什么意思| 甘油三酯高吃什么能降下来| 247是什么意思| 吃什么白细胞升的最快| 吃冰糖有什么好处和坏处| 为什么印度人叫阿三| 肉芽肿是什么病| pd是什么| 麒麟是什么| 铁锈用什么能洗掉| 硅胶是什么材质| 排骨炖什么补血补气| 什么的枣| 女人下身干燥无水是什么原因| 为什么得甲亢| 自勉是什么意思| 两特两重指的是什么| 胸闷气短是什么症状| 宫颈管少量积液是什么意思| 胃黏膜受损是什么症状| 解惑是什么意思| 梦到生儿子有什么预兆| 肝火是什么原因引起的| qd是什么意思| 冬虫夏草长什么样| 做梦抓到很多鱼是什么征兆| 天亮是什么时辰| 猪和什么属相最配| 经常手麻是什么原因引起的| 什么饼不能吃脑筋急转弯| 农历10月22日是什么星座| 肝功能是什么| aaa是什么意思| 农历5月是什么月| 湿热吃什么水果| 长江学者是什么级别| 汗蒸有什么好处| 八月七号是什么星座| 甘油是什么成分| 难以启齿是什么意思| 国防科技大学毕业是什么军衔| c5是什么驾驶证| 梦见家里办丧事是什么预兆| 吃什么清肺养肺| 丁克夫妻是什么意思| 晚上梦到蛇是什么意思| 离宅是什么意思| 低压高是什么原因造成的| 男人喜欢什么罩杯| 弯了是什么意思| 乳腺癌多发于什么年龄| 耳朵疼什么原因| 菜花是什么病| 房子风水主要看什么| 子宫薄是什么原因造成的| 肝脏在什么位置图片| 搞基什么意思| 什么是精神出轨| 什么的积雪| 十二月二十三是什么星座| 什么是工作| 五十知天命是什么意思| 遗尿是什么症状| 子宫内膜16mm说明什么| 黄金果是什么水果| 更年期吃什么好| 2025年是什么命| 抗日战争什么时候开始的| 夜晚睡不着觉什么原因| 黄芪的读音是什么| 白细胞数目偏高是什么意思| soie是什么面料| 前方高能什么意思| 女性膀胱炎吃什么药| 腿肿是什么原因| 蛇为什么怕雄黄| 傀儡什么意思| 碳足迹是什么| 甲钴胺的副作用是什么| 安乐死是什么| 倒嗓是什么意思| 查胆囊挂什么科| 青少年流鼻血是什么原因引起的| 聊是什么意思| 急性肠胃炎吃什么食物| 梅花是什么颜色的| 成吉思汗是什么族| 鳄鱼的天敌是什么| 血糖高吃什么| 海葡萄是什么| poscer是什么牌子的手表| 五月五日什么星座| 岳云鹏为什么这么火| 猫鼻支是什么症状| camouflage什么意思| 喜欢是什么感觉| 载脂蛋白b高是什么原因| 什么牌子的益生菌调理肠胃比较好| 张衡发明了什么东西| 一什么人家| 什么面什么方| 煮肉放什么调料| 热伤风吃什么| 侏儒症是什么原因引起的| 查处是什么意思| 曹操是什么星座| 775是什么意思| 鞭长莫及是什么意思| 蛇五行属什么| 6岁儿童为什么会长腿毛| 盗墓笔记它到底是什么| siv是什么意思| 11月11日是什么星座| 我拿什么留住你| 男才女貌是什么意思| 感冒黄鼻涕吃什么药| 醉酒当歌什么意思| 射手座和什么座最配对| 灯塔是什么意思| 失眠缺什么维生素| 心焦是什么意思| 肌肉僵硬是什么原因引起的| 什么样的女人水多| 梦见被猪咬是什么意思| 眼睛不好吃什么补眼睛| 牙周炎吃什么药效果好| 腿上有青筋是什么原因| 脚趾头抽筋是什么原因引起的| 后背疼去医院挂什么科| 什么样的血管瘤不用治| 犹太是什么意思| 6代表什么| 2001年是什么年| 脾胃寒湿吃什么中成药| 女人经常喝什么汤养颜| 禁忌什么意思| 乌龟和鳖有什么区别| 交公粮是什么意思| 粘人是什么意思| 什么时间最容易受孕| 怀孕六个月出血是什么原因| 先天愚型是什么病| 薛之谦为什么离婚| 脚褪皮是什么原因| 不复相见什么意思| 世界七大奇迹分别是什么| 梦见自己鼻子流血是什么预兆| 牙膏尾部的颜色代表什么意思| 尿等待吃什么药最好| 通讯地址填什么| 嬴稷和嬴政什么关系| 男性性功能减退吃什么药| 陶氏腔积液是什么意思| 吃什么补营养最快| 胃窦黄斑瘤是什么病| 为什么流褐色分泌物| bug是什么意思中文翻译| 仓鼠吃什么食物| 旅游的意义是什么| 喝最烈的酒下一句是什么| 妈妈是什么意思| 抗组胺是什么意思| 空虚是什么意思| 肠胃炎吃什么食物好| 人格分裂什么意思| 外阴炎吃什么药| tin是什么| 宁字五行属什么的| 哈达是什么意思| 石花菜是什么植物| 突然腰疼是什么原因| 腿浮肿是什么原因引起的| 男士感染霉菌用什么药| tony是什么意思| 手脚热是什么原因| 桃花什么颜色| 说女人强势是什么意思| 什么是提肛运动| 脚抽筋是什么原因引起的| 降血脂吃什么药效果好| baleno是什么牌子| 什么是虚无主义| 嗓子沙哑吃什么药| 痛风什么东西不可以吃| 前列腺增大伴钙化是什么意思| 施华蔻属于什么档次| 陶白白是什么星座| 苹果熬水喝有什么功效| 女人的第二张脸是什么| dn是什么意思| 物质是由什么组成的| 酸中毒是什么意思| 沙参长什么样子图片| 什么是热辐射| 凝望什么| 梦见捡鸡蛋是什么意思| 光天化日什么意思| 浙江有什么特产| 脑缺血有什么症状| 酱油什么时候发明的| 月经提前是什么原因引起的| 空挡是什么意思| 视力矫正是什么意思| 排异是什么意思| 水油是什么| 孩子专注力差去什么医院检查| 颈部淋巴结肿大吃什么药| 牛津布是什么材质| 什么粥减肥效果好| 红颜什么意思| 粤语点解什么意思| 百度Jump to content

《阴阳师》绿色度测评报告

From Wikipedia, the free encyclopedia
百度 要结合中纪委查处的安徽省“陈树隆、杨振超、周春雨案”以及省部分领导干部严重违法违纪案例,教育警示党员干部受警醒、知敬畏、明底线。

In computer science, reference counting is a programming technique of storing the number of references, pointers, or handles to a resource, such as an object, a block of memory, disk space, and others.

In garbage collection algorithms, reference counts may be used to deallocate objects that are no longer needed.

Advantages and disadvantages

[edit]

The main advantage of the reference counting over tracing garbage collection is that objects are reclaimed as soon as they can no longer be referenced, and in an incremental fashion, without long pauses for collection cycles and with clearly defined lifetime of every object. In real-time applications or systems with limited memory, this is important to maintain responsiveness. Reference counting is also among the simplest forms of memory management to implement. It also allows for effective management of non-memory resources such as operating system objects, which are often much scarcer than memory (tracing garbage collection systems use finalizers for this,[citation needed] but the delayed reclamation may cause problems). Weighted reference counts are a good solution for garbage collecting a distributed system.

Circular list example from a 1985 Master's Thesis.[1] Rectangles denote cons pairs, with reference counts. Even if the incoming top left pointer is removed, all counts remain >0.

Tracing garbage collection cycles are triggered too often if the set of live objects fills most of the available memory;[citation needed] it requires extra space to be efficient.[citation needed] Reference counting performance does not deteriorate as the total amount of free space decreases.[2]

Reference counts are also useful information to use as input to other runtime optimizations. For example, systems that depend heavily on immutable objects such as many functional programming languages can suffer an efficiency penalty due to frequent copies.[citation needed] However, if the compiler (or runtime system) knows that a particular object has only one reference (as most do in many systems), and that the reference is lost at the same time that a similar new object is created (as in the string append statement str ← str + "a"), it can replace the operation with a mutation on the original object.

Reference counting in naive form has three main disadvantages over the tracing garbage collection, both of which require additional mechanisms to ameliorate:

  • The frequent updates it involves are a source of inefficiency. While tracing garbage collectors can impact efficiency severely via context switching and cache line faults, they collect relatively infrequently, while accessing objects is done continually. Also, less importantly, reference counting requires every memory-managed object to reserve space for a reference count. In tracing garbage collectors, this information is stored implicitly in the references that refer to that object, saving space, although tracing garbage collectors, particularly incremental ones, can require additional space for other purposes.
  • The naive algorithm described above can't handle reference cycles, an object which refers directly or indirectly to itself. A mechanism relying purely on reference counts will never consider cyclic chains of objects for deletion, since their reference count is guaranteed to stay nonzero (cf. picture). Methods for dealing with this issue exist but can also increase the overhead and complexity of reference counting — on the other hand, these methods need only be applied to data that might form cycles, often a small subset of all data. One such method is the use of weak references, while another involves using a mark-sweep algorithm that gets called infrequently to clean up.
  • In a concurrent setting, all updates of the reference counts and all pointer modifications must be atomic operations, which incurs an additional cost. There are three reasons for the atomicity requirements. First, a reference count field may be updated by multiple threads, and so an adequate atomic instruction, such as a (costly) compare-and-swap, must be used to update the counts. Second, it must be clear which object loses a reference so that its reference count can be adequately decremented. But determining this object is non-trivial in a setting where multiple threads attempt to modify the same reference (i.e., when data races are possible). Finally, there exists a subtle race in which one thread gains a pointer to an object, but before it increments the object's reference count, all other references to this object are deleted concurrently by other threads and the object is reclaimed, causing the said thread to increment a reference count of a reclaimed object.

In addition to these, if the memory is allocated from a free list, reference counting suffers from poor locality. Reference counting alone cannot move objects to improve cache performance, so high performance collectors implement a tracing garbage collector as well. Most implementations (such as the ones in PHP and Objective-C) suffer from poor cache performance since they do not implement copying objects.[3]

Graph interpretation

[edit]

When dealing with garbage collection schemes, it is often helpful to think of the reference graph, which is a directed graph where the vertices are objects and there is an edge from an object A to an object B if A holds a reference to B. We also have a special vertex or vertices representing the local variables and references held by the runtime system, and no edges ever go to these nodes, although edges can go from them to other nodes.

In this context, the simple reference count of an object is the in-degree of its vertex. Deleting a vertex is like collecting an object. It can only be done when the vertex has no incoming edges, so it does not affect the out-degree of any other vertices, but it can affect the in-degree of other vertices, causing their corresponding objects to be collected as well if their in-degree also becomes 0 as a result.

The connected component containing the special vertex contains the objects that can't be collected, while other connected components of the graph only contain garbage. If a reference-counting garbage collection algorithm is implemented, then each of these garbage components must contain at least one cycle; otherwise, they would have been collected as soon as their reference count (i.e., the number of incoming edges) dropped to zero.

Dealing with inefficiency of updates

[edit]

Incrementing and decrementing reference counts every time a reference is created or destroyed can significantly impede performance. Not only do the operations take time, but they damage cache performance and can lead to pipeline bubbles. Even read-only operations like calculating the length of a list require a large number of reads and writes for reference updates with naive reference counting.

One simple technique is for the compiler to combine a number of nearby reference updates into one. This is especially effective for references which are created and quickly destroyed. Care must be taken, however, to put the combined update at the right position so that a premature free can be avoided.

The Deutsch-Bobrow method of reference counting capitalizes on the fact that most reference count updates are in fact generated by references stored in local variables. It ignores these references, only counting references in data structures, but before an object with reference count zero can be deleted, the system must verify with a scan of the stack and registers that no other reference to it still exists.

Another technique devised by Henry Baker involves deferred increments,[4] in which references which are stored in local variables do not immediately increment the corresponding reference count, but instead defer this until it is necessary. If such a reference is destroyed quickly, then there is no need to update the counter. This eliminates a large number of updates associated with short-lived references (such as the above list-length-counting example). However, if such a reference is copied into a data structure, then the deferred increment must be performed at that time. It is also critical to perform the deferred increment before the object's count drops to zero, to avoid a premature free.

A dramatic decrease in the overhead on counter updates was obtained by Levanoni and Petrank.[5][6] They introduce the update coalescing method which coalesces many of the redundant reference count updates. Consider a pointer that in a given interval of the execution is updated several times. It first points to an object O1, then to an object O2, and so forth until at the end of the interval it points to some object On. A reference counting algorithm would typically execute rc(O1)--, rc(O2)++, rc(O2)--, rc(O3)++, rc(O3)--, ..., rc(On)++. But most of these updates are redundant. In order to have the reference count properly evaluated at the end of the interval it is enough to perform rc(O1)-- and rc(On)++. The rest of the updates are redundant.

Levanoni and Petrank showed in 2001 how to use such update coalescing in a reference counting collector. When using update coalescing with an appropriate treatment of new objects, more than 99% of the counter updates are eliminated for typical Java benchmarks.

Interestingly, update coalescing also eliminates the need to employ atomic operations during pointer updates in a concurrent setting, this solving reference counting issues in a concurrent setting. Therefore, update coalescing solves the third problem of naive reference counting (i.e., a costly overhead in a concurrent setting). Levanoni and Petrank presented an enhanced algorithm that may run concurrently with multithreaded applications employing only fine synchronization.[7]

Blackburn and McKinley's ulterior reference counting method in 2003[8] combines deferred reference counting with a copying nursery, observing that the majority of pointer mutations occur in young objects. This algorithm achieves throughput comparable with the fastest generational copying collectors with the low bounded pause times of reference counting.

Dealing with reference cycles

[edit]

Perhaps the most obvious way to handle reference cycles is to design the system to avoid creating them. A system may explicitly forbid reference cycles; file systems with hard links often do this. Judicious use of "weak" (non-counted) references may also help avoid retain cycles; the Cocoa framework, for instance, recommends using "strong" references for parent-to-child relationships and "weak" references for child-to-parent relationships.[9]

Systems may also be designed to tolerate or correct the cycles they create in some way. Developers may design code to explicitly "tear down" the references in a data structure when it is no longer needed, though this has the cost of requiring them to manually track that data structure's lifetime. This technique can be automated by creating an "owner" object that does the tearing-down when it is destroyed; for instance, a Graph object's destructor could delete the edges of its GraphNodes, breaking the reference cycles in the graph. Cycles may even be ignored in systems with short lives and a small amount of cyclic garbage, particularly when the system was developed using a methodology of avoiding cyclic data structures wherever possible, typically at the expense of efficiency.

Computer scientists have also discovered ways to detect and collect reference cycles automatically, without requiring changes in the data structure design. One simple solution is to periodically use a tracing garbage collector to reclaim cycles; since cycles typically constitute a relatively small amount of reclaimed space, the collector can be run much less often than with an ordinary tracing garbage collector.

Bacon describes a cycle-collection algorithm for reference counting with similarities to tracing collectors, including the same theoretical time bounds. It is based on the observation that a cycle can only be isolated when a reference count is decremented to a nonzero value. All objects which this occurs on are put on a roots list, and then periodically the program searches through the objects reachable from the roots for cycles. It knows it has found a cycle that can be collected when decrementing all the reference counts on a cycle of references brings them all down to zero.[10] An enhanced version of this algorithm by Paz et al.[11] is able to run concurrently with other operations and improve its efficiency by using the update coalescing method of Levanoni and Petrank.[5][6]

Variant forms

[edit]

Although it is possible to augment simple reference counts in a variety of ways, often a better solution can be found by performing reference counting in a fundamentally different way. Here we describe some of the variants on reference counting and their benefits and drawbacks.

Weighted reference counting

[edit]

In weighted reference counting, each reference is assigned a weight, and each object tracks not the number of references referring to it, but the total weight of the references referring to it. The initial reference to a newly created object has a large weight, such as 216. Whenever this reference is copied, half of the weight goes to the new reference, and half of the weight stays with the old reference. Since the total weight does not change, the object's reference count does not need to be updated.

Destroying a reference decrements the total weight by the weight of that reference. When the total weight becomes zero, all references have been destroyed. If an attempt is made to copy a reference with a weight of 1, the reference has to "get more weight" by adding to the total weight and then adding this new weight to the reference, and then splitting it. An alternative in this situation is to create an indirection reference object, the initial reference to which is created with a large weight which can then be split.

The property of not needing to access a reference count when a reference is copied is particularly helpful when the object's reference count is expensive to access, for example because it is in another process, on disk, or even across a network. It can also help increase concurrency by avoiding many threads locking a reference count to increase it. Thus, weighted reference counting is most useful in parallel, multiprocess, database, or distributed applications.

The primary problem with simple weighted reference counting is that destroying a reference still requires accessing the reference count, and if many references are destroyed, this can cause the same bottlenecks we seek to avoid. Some adaptations of weighted reference counting seek to avoid this by transferring weight from a dying reference to an active reference.

Weighted reference counting was independently devised by Bevan[12] and Watson & Watson[13] in 1987.

Indirect reference counting

[edit]

In indirect reference counting, it is necessary to keep track of the reference's source. This means that two references are kept to the object: a direct one which is used for invocations; and an indirect one which forms part of a diffusion tree, such as in the Dijkstra–Scholten algorithm, which allows a garbage collector to identify dead objects. This approach prevents an object from being discarded prematurely.

Examples of use

[edit]

Garbage collection

[edit]

As a collection algorithm, reference counting tracks, for each object, a count of the number of references to it held by other objects. If an object's reference count reaches zero, the object has become inaccessible, and can be destroyed.

When an object is destroyed, any objects referenced by that object also have their reference counts decreased. Because of this, removing a single reference can potentially lead to a large number of objects being freed. A common modification allows reference counting to be made incremental: instead of destroying an object as soon as its reference count becomes zero, it is added to a list of unreferenced objects, and periodically (or as needed) one or more items from this list are destroyed.

Simple reference counts require frequent updates. Whenever a reference is destroyed or overwritten, the reference count of the object it references is decremented, and whenever one is created or copied, the reference count of the object it references is incremented.

Reference counting is also used in file systems and distributed systems, where full non-incremental tracing garbage collection is too time-consuming because of the size of the object graph and slow access speed.[14]

Component Object Model

[edit]

Microsoft's Component Object Model (COM) and WinRT makes pervasive use of reference counting. In fact, two of the three methods that all COM objects must provide (in the IUnknown interface) increment or decrement the reference count. Much of the Windows Shell and many Windows applications (including MS Internet Explorer, MS Office, and countless third-party products) are built on COM, demonstrating the viability of reference counting in large-scale systems.[citation needed]

One primary motivation for reference counting in COM is to enable interoperability across different programming languages and runtime systems. A client need only know how to invoke object methods in order to manage object life cycle; thus, the client is completely abstracted from whatever memory allocator the implementation of the COM object uses. As a typical example, a Visual Basic program using a COM object is agnostic towards whether that object was allocated (and must later be deallocated) by a C++ allocator or another Visual Basic component.

C++

[edit]

C++ does not perform reference-counting by default, fulfilling its philosophy of not adding functionality that might incur overheads where the user has not explicitly requested it. Objects that are shared but not owned can be accessed via a reference, raw pointer, or iterator (a conceptual generalisation of pointers).

However, by the same token, C++ provides native ways for users to opt-into such functionality: C++11 provides reference counted smart pointers, via the std::shared_ptr class, enabling automatic shared memory-management of dynamically allocated objects. Programmers can use this in conjunction with weak pointers (via std::weak_ptr) to break cyclic dependencies. Objects that are dynamically allocated but not intended to be shared can have their lifetime automatically managed using a std::unique_ptr.

In addition, C++11's move semantics further reduce the extent to which reference counts need to be modified by removing the deep copy normally used when a function returns an object, as it allows for a simple copy of the pointer of said object.

Cocoa (Objective-C)

[edit]

Apple's Cocoa and Cocoa Touch frameworks (and related frameworks, such as Core Foundation) use manual reference counting, much like COM. Traditionally this was accomplished by the programmer manually sending retain and release messages to objects, but Automatic Reference Counting, a Clang compiler feature that automatically inserts these messages as needed, was added in iOS 5[15] and Mac OS X 10.7.[16] Mac OS X 10.5 introduced a tracing garbage collector as an alternative to reference counting, but it was deprecated in OS X 10.8 and removed from the Objective-C runtime library in macOS Sierra.[17][18] iOS has never supported a tracing garbage collector.

Delphi

[edit]

Delphi is mostly not a garbage collected language, in that user-defined types must still be manually allocated and deallocated; however, it does provide automatic collection using reference counting for a few built-in types, such as strings, dynamic arrays, and interfaces, for ease of use and to simplify the generic database functionality. It is up to the programmer to decide whether to use the built-in types; Delphi programmers have complete access to low-level memory management like in C/C++. So all potential cost of Delphi's reference counting can, if desired, be easily circumvented.

Some of the reasons reference counting may have been preferred to other forms of garbage collection in Delphi include:

  • The general benefits of reference counting, such as prompt collection.
  • Cycles either cannot occur or do not occur in practice because none of the garbage-collected built-in types are recursive. (using interfaces one could create such scenario, but that is not common usage)
  • The overhead in code size required for reference counting is very small (on native x86, typically a single LOCK INC, LOCK DEC or LOCK XADD instruction, which ensures atomicity in any environment), and no separate thread of control is needed for collection as would be needed for a tracing garbage collector.
  • Many instances of the most commonly used garbage-collected type, the string, have a short lifetime, since they are typically intermediate values in string manipulation. A lot of local string usage could be optimized away, but the compiler currently doesn't do it.
  • The reference count of a string is checked before mutating a string. This allows reference count 1 strings to be mutated directly whilst higher reference count strings are copied before mutation. This allows the general behaviour of old style pascal strings to be preserved whilst eliminating the cost of copying the string on every assignment.
  • Because garbage-collection is only done on built-in types, reference counting can be efficiently integrated into the library routines used to manipulate each datatype, keeping the overhead needed for updating of reference counts low. Moreover, a lot of the runtime library is in hand-optimized assembler.
  • The string type can be cast to a pointer to char, and high performance operations can be performed that way. This is important since both Delphi and FPC implement their RTL in Pascal. Various other automated types have such casting options.

GObject

[edit]

The GObject object-oriented programming framework implements reference counting on its base types, including weak references. Reference incrementing and decrementing uses atomic operations for thread safety. A significant amount of the work in writing bindings to GObject from high-level languages lies in adapting GObject reference counting to work with the language's own memory management system.

The Vala programming language uses GObject reference counting as its primary garbage collection system, along with copy-heavy string handling.[19]

Perl

[edit]

Perl also uses reference counting, without any special handling of circular references, although (as in Cocoa and C++ above), Perl does support weak references, which allows programmers to avoid creating a cycle.

PHP

[edit]

PHP uses a reference counting mechanism for its internal variable management.[20] Since PHP 5.3, it implements the algorithm from Bacon's above mentioned paper. PHP allows you to turn on and off the cycle collection with user-level functions. It also allows you to manually force the purging mechanism to be run.

Python

[edit]

Python also uses reference counting and offers cycle detection as well (and can reclaim reference cycles).[21]

Rust

[edit]

Like other low-level languages, Rust does not provide reference counting by default. Instead, any constructed type will be dropped when it falls out of scope. When a programmer needs to define the scope of a constructed type, they often use lifetimes.

However, the language also offers various alternatives to complex forms of memory management. Reference counting functionality is provided by the Rc and Arc types, which are non-atomic and atomic respectively.

For example, the type Rc<T> provides shared ownership of a value of type T, allocated on the heap for multiple references to its data.[22]

use std::rc::Rc;

struct Cat {
    color: String,
}

fn main() {
    let cat = Cat { color: "black".to_string() };
    let cat = Rc::new(cat);
}

Using these constructs allows programmers to avoid lifetimes for a small runtime cost. Both reference counters keep track of the number of owners, as they must drop themselves when no owners remain.

One noteworthy facet of these types is related to their usage as a shared reference. In Rust, shared references cannot mutate their held data, so Rc often comes bundled with Cell, and Arc with Mutex, in contexts where interior mutability is necessary.

Interior mutability without UnsafeCell has performance costs, too, so, for maximum performance, some applications may call for additional complexity.[23]

Squirrel

[edit]

Squirrel uses reference counting with cycle detection. This tiny language is relatively unknown outside the video game industry; however, it is a concrete example of how reference counting can be practical and efficient (especially in realtime environments).[citation needed]

Swift

[edit]

Swift uses reference counting to track and manage the memory of class instances, and provides the weak keyword for creating weak references. Instances of value types do not use reference counting.[24]

Tcl

[edit]

Tcl 8 uses reference counting for memory management of values (Tcl Obj structs). Since Tcl's values are immutable, reference cycles are impossible to form and no cycle detection scheme is needed. Operations that would replace a value with a modified copy are generally optimized to instead modify the original when its reference count indicates that it is not shared. The references are counted at a data structure level, so the problems with very frequent updates discussed above do not arise.

Xojo

[edit]

Xojo also uses reference counting, without any special handling of circular references, although (as in Cocoa and C++ above), Xojo does support weak references, which allows programmers to avoid creating a cycle.

File systems

[edit]

Many file systems maintain reference counts to any particular block or file, for example the inode link count on Unix-style file systems, which are usually known as hard links. When the count reaches zero, the file can be safely deallocated. While references can still be made from directories, some Unixes only allow references from live processes, and there can be files that exist outside the file system hierarchy.

References

[edit]
  1. ^ Kevin G. Cassidy (December 1985). The Feasibility of Automatic Storage Reclamation with Concurrent Program Execution in a LISP Environment (PDF) (Master's thesis). Naval Postgraduate School, Monterey/CA. Here: p.25
  2. ^ Wilson, Paul R. (1992). "Uniprocessor Garbage Collection Techniques". Proceedings of the International Workshop on Memory Management. London, UK: Springer-Verlag. pp. 1–42. ISBN 3-540-55940-X. Section 2.1.
  3. ^ Rifat Shahriyar, Stephen M. Blackburn, Xi Yang and Kathryn S. McKinley (2013). "Taking Off the Gloves with Reference Counting Immix" (PDF). 24th ACM SIGPLAN conference on Object Oriented Programming Systems, Languages and Applications. OOPSLA 2013. doi:10.1145/2509136.2509527.{{cite conference}}: CS1 maint: multiple names: authors list (link)
  4. ^ Henry Baker (September 1994). "Minimizing Reference Count Updating with Deferred and Anchored Pointers for Functional Data Structures". ACM SIGPLAN Notices. 29 (9): 38–43. CiteSeerX 10.1.1.25.955. doi:10.1145/185009.185016. S2CID 14448488.
  5. ^ a b Yossi Levanoni, Erez Petrank (2001). "An on-the-fly reference-counting garbage collector for java". Proceedings of the 16th ACM SIGPLAN conference on Object-oriented programming, systems, languages, and applications. OOPSLA 2001. pp. 367–380. doi:10.1145/504282.504309.
  6. ^ a b Yossi Levanoni, Erez Petrank (2006). "An on-the-fly reference-counting garbage collector for java". ACM Trans. Program. Lang. Syst. 28: 31–69. CiteSeerX 10.1.1.15.9106. doi:10.1145/1111596.1111597. S2CID 14777709.
  7. ^ "An On-the-Fly Reference-Counting Garbage Collector for Java" (PDF). Cs.technion.ac.il. Retrieved 24 June 2017.
  8. ^ Stephen Blackburn; Kathryn McKinley (2003). "Ulterior Reference Counting: Fast Garbage Collection without a Long Wait" (PDF). Proceedings of the 18th annual ACM SIGPLAN conference on Object-oriented programing, systems, languages, and applications. OOPSLA 2003. pp. 344–358. doi:10.1145/949305.949336. ISBN 1-58113-712-5.
  9. ^ "Mac Developer Library". Developer.apple.com. Retrieved 17 December 2015.
  10. ^ Bacon, David F.; Rajan, V. T. (2001). "Concurrent Cycle Collection in Reference Counted Systems" (PDF). ECOOP 2001 — Object-Oriented Programming. Lecture Notes in Computer Science. Vol. 2072. pp. 207–235. doi:10.1007/3-540-45337-7_12. ISBN 978-3-540-42206-8. Archived from the original (PDF) on 23 July 2004.
  11. ^ Harel Paz, David F. Bacon, Elliot K. Kolodner, Erez Petrank, V. T. Rajan (2007). "An efficient on-the-fly cycle collection". ACM Transactions on Programming Languages and Systems. 29 (4): 20–es. CiteSeerX 10.1.1.10.2777. doi:10.1145/1255450.1255453. S2CID 4550008.{{cite journal}}: CS1 maint: multiple names: authors list (link)
  12. ^ Bevan, D. I. (1987). "Distributed garbage collection using reference counting". Volume II: Parallel Languages on PARLE: Parallel Architectures and Languages Europe. Eindhoven, The Netherlands: Springer-Verlag. pp. 176–187. ISBN 0-387-17945-3.
  13. ^ Watson, Paul; Watson, Ian (1987). "An efficient garbage collection scheme for parallel computer architectures". Volume II: Parallel Languages on PARLE: Parallel Architectures and Languages Europe. Eindhoven, The Netherlands: Springer-Verlag. pp. 432–443. ISBN 0-387-17945-3.
  14. ^ Bruno, Rodrigo; Ferreira, Paulo (2018). "A Study on Garbage Collection Algorithms for Big Data Environments". ACM Computing Surveys. 51: 1–35. doi:10.1145/3156818. S2CID 21388487.
  15. ^ [1] Archived 9 June 2011 at the Wayback Machine
  16. ^ "Mac Developer Library". Developer.apple.com. Retrieved 17 December 2015.
  17. ^ Siracusa, John (25 July 2012). "OS X 10.8 Mountain Lion: the Ars Technica review". Ars Technica. At section "Objective-C enhancements". Retrieved 17 November 2016.
  18. ^ "Xcode 8 Release Notes". Apple Developer. 27 October 2016. Archived from the original on 19 March 2017. Retrieved 19 March 2017.
  19. ^ "Projects/Vala/ReferenceHandling - GNOME Wiki!". GNOME. 25 May 2015. Retrieved 17 December 2015.
  20. ^ "PHP: Reference Counting Basics - Manual". www.php.net. Retrieved 1 October 2020.
  21. ^ "1. Extending Python with C or C++ — Python 2.7.11 documentation". Docs.python.org. 5 December 2015. Retrieved 17 December 2015.
  22. ^ "std::rc - Rust". doc.rust-lang.org. Retrieved 2 November 2020.
  23. ^ "The Rust Reference". 21 July 2022. Interior Mutability. Archived from the original on 24 March 2024. Retrieved 22 April 2024.
  24. ^ "Documentation". docs.swift.org. Retrieved 6 December 2023.
[edit]
男人眉骨高代表什么 冲鼠煞北是什么意思 烤冷面的面皮是什么面 生理曲度存在是什么意思 甲状腺挂什么科
着床是什么意思 可乐定是什么药 碱性食物对身体有什么好处 狗为什么不吃饭 吃狗肉不能和什么一起吃
全身发烫但不发烧是什么原因 日本浪人是什么意思 肌肉痉挛用什么药能治好 截瘫是什么意思 gree是什么牌子
属马女和什么属相最配 什么叫蛇胆疮 tony是什么意思 荔枝和什么不能一起吃 心包隐窝是什么意思
灼热感是什么样的感觉hcv9jop1ns2r.cn lo娘是什么意思hcv7jop5ns3r.cn 扁桃体发炎不能吃什么hcv8jop8ns7r.cn 乙型肝炎表面抗原阳性是什么意思hcv8jop9ns5r.cn 龙飞凤舞是什么意思hcv7jop4ns5r.cn
母仪天下是什么意思hcv8jop8ns6r.cn 屁股出血是什么原因hcv9jop3ns5r.cn 什么是问题shenchushe.com 鹅蛋什么人不能吃zsyouku.com 369是什么意思hcv8jop8ns0r.cn
甲亢病是什么病hcv7jop9ns4r.cn 双肾囊性灶是什么意思hcv9jop8ns1r.cn 百年老枞属于什么茶hebeidezhi.com 小孩肺热吃什么好清肺热hcv9jop1ns0r.cn 尿酸高吃什么药aiwuzhiyu.com
岁寒三友指什么hcv8jop8ns9r.cn toryburch什么牌子wzqsfys.com 人不出汗是什么原因hcv8jop6ns9r.cn 子字属于五行属什么aiwuzhiyu.com 手信是什么hcv7jop9ns7r.cn
百度