- 论坛徽章:
- 0
|
![]()
3.6.1 Per-upcall event ordering 回调前的事件次序
The techniques described in the previous section for storing events separate the event types that pass TCBs from those that don't. This means that event ordering is not implicit, and would require extra work to preserve. However, it turns out that in most cases, order doesn't matter, and in the cases that it does matter, the UTS can always determine order.
前面的章节描述了存储事件,区分事件类型的技术忽略了TCB来自那些禁止事项。这意味着事件的顺序不是固定的,可能需要额外的工作去保护。但是在大多数情况下,顺序没关系,即使在顺序有关系的情况下,UTS总能决定顺序
As an example, suppose thread_block() and thread_unblock() events occur for the same Thread ID in the same upcall. In one interpretation, where the thread_block() event occurred first, the UTS last knew the corresponding thread to be running, so the two events simply mean that the thread can be continued. In the other interpretation, where the thread_unblock() event occurred first, the Thread ID has been recycled. However, this isn't possible, since there must have been an intervening thread_new() event, which couldn't have happened until after the Thread unblocked.
举个例子,假设一个线程在一次回调是thread_block()和thread_unblock()事件都发生了。一种情况,如果thread_block()先发生,UTS最后知道通信的线程在运行,所以两个事件表示该线程可以继续执行。另一种情况,如果thread_unblock事件先发生,线程的ID可以被重新利用,但是,这不可能,因为它一定是一个在thread_new()事件之间,在一个线程被解锁之前是不可能发生的。
3.7 Upcall parallelism 回调对应
This section is not yet adequately fleshed out. Issues to consider:
这部分还有没适当的充实,关键要考虑
How do upcalls from multiple processors keep from stomping on each other? For example, if we have only one chain of TCBs, how does it get safely used?
怎么才能避免回调在多个处理器上跳来跳去,比如,如果只有一链TCB,怎么能让它被安全的使用。
What if the UTS doesn't process the TCBs before another upcall happens? Do we run the risk of overflowing the TCB chain?
如果在新的回调发生时,UTS不能处理TCB怎么办,要运行这个存在溢出危险的TCB链吗
* How do we handle races between the UTS and the kernel? For example, signals cannot be cleared until delivered to a thread, and the kernel must look at the userland bitmap to see if a signal is still pending.
怎么处理UTS和内核间的竞争呢,比如,在信号传递给线程前不能被清空,那么内核必须检查用户空间的位图,去看是否信号还在传递中。
3.8 Kernel scheduler 内核调度
The kernel scheduler needs a major overhaul in order to support KSEs. Support for the following features is necessary:
内核调度需要进行仔细检查以便支持KSE,下面的特征是必须支持的
* Soft affinity. The scheduler should try to keep processes running on the same processors if there is the possibility of reaping benefits from increased cache locality.
软亲缘。调度程序应该尽力保证一个进程在被调度前后运行在同一个CPU 上,这样可以通过增加缓存的击中率提高性能。
* Hard affinity (binding). Some programs may wish to bind a KSE (pedantically speaking, a KSEG is bound) to a particular processor in order to improve cache locality, or to tighten the bounds on a real-time task. Bound KSEs are run only on one CPU, unless priority inheritance forces it to be run on another CPU in order to avoid deadlock due to priority inversion.
硬亲缘(绑定)。一些程序可能希望把KSE绑定在一个单独的CPU上(准确的说,KSEG)来提高缓存的击中率,或是让实时任务更严密,被绑定的KSE只运行在一个CPU上,除非在优先级倒置时,为了避免死锁,优先级继承迫使它运行在另一个CPU上。
* Increased parallelism. The current scheduler can only be executed by one processor at a time, which makes it a severe bottleneck, depending on the type of system load.
对应增长。在同一时刻内,调度程序只能运行在一个处理器上,随着系统的负载量增大,这产生了严重的瓶颈。
Part of the solution is to implement per-CPU scheduling queues. Each CPU has an associated queue for softly affined KSEs, as well as a queue for bound KSEs. In addition, there is a global queue for fixed-priority KSEs, such as interrupt threads and real-time KSEs. Fixed-priority KSEs prefer to run on the same CPU as the last time they were run, but scheduling issues keep more complex affinity algorithms from being beneficial for this class of KSEs. See Figure 9 for a simplistic representation of the various scheduling queues. The diagram ignores some details such as split queues for various priorities.
这个问题的解决方案是实现单CPU调度队列。每个CPU有一个软亲缘KSE队列和一个绑定KSE队列。此外,还有一个为固定优先级KSE设的全局队列。比如中断处理线程和实时KSE。确定优先级的KSE更希望运行在他们上次运行的CPU上,但是调度算法有更多对这类KSE有益的负杂亲缘算法。图9显示了一个十分简单的各种调度队列的代表。图忽略了某些细节,比如把队列分成不同优先级
Each CPU schedules from its own queues, and resorts to stealing runnable softly affined KSEs from other CPUs if there are no runnable KSEs. Every so often (exact number to be determined by benchmarking, but a likely number is 5 seconds), CPU loads are calculated, and if the loads are too disparate, the under-loaded CPU(s) steal additional KSEs from the overloaded CPU(s) in an attempt to balance the load. Instantaneous per-CPU load is defined as the number of runnable timesharing KSEs in a CPU's queues. Instantaneous system load is the sum of the instantaneous per-CPU loads. Note that normal (non-balancing) KSE stealing is adequate to keep the system busy if there is enough work to do, but that without periodic load balancing, time sharing becomes unfair if there aren't approximately equal CPU loads.
每个CPU在自己的队列中进行调度,如果没有可运行的KSE,则从其他CPU那里取一些有软亲缘的KSE,并且重新排序。每过一段时间(确切得数子通过测试决定,但是可能的是5秒)会计算CPU负载,如果CPU之间的负载差别很大低负载的CPU从高负载CPU的调度队列里取一些KSE加入自己的队列,以达到负载平衡。瞬间单CPU负载定义为在当前CPU的队列里,可运行的分时KSE的数量。瞬间系统负载是瞬间单CPU负载的和。注意,正常的KSE获取是适量的,为了保证在有足够工作时,系统的繁忙。如果不进行周期性的负载平衡,如果没有大致均衡的CPU负载,分时就会变得不公平。
This is essentially the approach that was taken in DEC OSF/1 3.0 [Denham], and it seems to have worked well there. One notable difference between our kernel and OSF/1 is that our interrupts are backed by threads, whereas OSF/1 keeps spl()s (IPLs in their terminology). However, this doesn't affect the design, since OSF/1 already has to handle real-time scheduling.
这个在本质上和DEC 的OSF/1.3.0相似,这种方法在OSF上好象很有效,一个值得注意的不同是,我们的中断是基于线程的,虽然OSF/1拥有spl,但是,这个没有影响设计,因为OSF/1必须处理实时调度。
Figure 9: Kernel scheduler queues 内核调度队列
![]()
4 Summary
Discussion about improved threads support in FreeBSD has been ongoing for several years. The KSE project aims to implement a kernel mechanism that allows the kernel and userland to support threaded processes and communicate with each other effectively so that the necessary information is available for both to do their jobs efficiently and correctly.
讨论在FreeBSD中增加线程支持已经进行了几年。KSE项目的目标是实现一个内核结构,允许内核和用户空间支持线程处理以及相互间有效的通信以便它们有效和正确的完成工作所必要的信息对两者来说都是可用的。
Glossary
KSE:
Kernel-scheduled entity. This gets scheduled.
Thread:
A runnable context. This gets suspended.
KSEG:
Kernel-scheduled entity group.
PCB:
Process control block.
SA:
Scheduler activation.
TCB:
Thread control block.
UTS:
Userland threads scheduler. Userland, multi-level, and scheduler activation-based threads libraries all have a UTS.
Bibliography
Anderson
Thomas E. Anderson, Brian N. Bershad, Edward D. Lazowska, and Henry M. Levy, Scheduler Activations: Effective Kernel Support for the User-Level Management of Parallelism, ACM Transactions on Computer Systems, Vol. 10, No. 1, February 1992, Pages 53-79.
Boykin
Joseph Boykin, David Kirschen, Alan Langerman, and Susan LoVerso, Programming under Mach, Addison-Wesley Publishing Company, Inc. (1993).
Butenhof
David R. Butenhof, Programming with POSIX threads, Addison Wesley Longman, Inc. (1997).
Denham
Jeffrey M. Denham, Paula Long, and James A. Woodward, DEC OSF/1 Symmetric Multiprocessing, Digital Technical Journal, Vol. 6, No. 3.
Kleiman
Steve Kleiman, Devang Shah, and Bart Smaalders, Programming with Threads, SunSoft Press (1996).
Lemon
Jonathan Lemon, Kqueue: A generic and scalable event notification facility, BSDcon Conference Proceedings (2000), http://people.freebsd.org/~jlemon/kqueue.pdf.
Mauro
Jim Mauro and Richard McDougall, Solaris Internals, Sun Microsystems Press (2001).
McKusick
Marshall Kirk McKusick, Keith Bostic, Michael J. Karels, and John S. Quarterman, The Design and Implementation of the 4.4BSD Operating System, Addison-Wesley Publishing Company, Inc. (1996).
Vahalia
Uresh Vahalia, UNIX $^{TM}$ Internals: The New Frontiers, Prentice-Hall, Inc. (1996).
About this document ...
Kernel-Scheduled Entities for FreeBSD
This document was generated using the LaTeX2HTML translator Version 2002-2-1 (1.70)
Copyright © 1993, 1994, 1995, 1996, Nikos Drakos, Computer Based Learning Unit, University of Leeds.
Copyright © 1997, 1998, 1999, Ross Moore, Mathematics Department, Macquarie University, Sydney.
The command line arguments were:
latex2html -split 0 -dir html -no_navigation -no_white -show_section_numbers freebsd_kse.tex
The translation was initiated by Chris Knight on 2003-01-01
Chris Knight 2003-01-01
[ 本帖最后由 DarkBlueSea 于 2006-4-10 13:11 编辑 ] |
|