core: Decompose task into more submodules
This commit is contained in:
parent
591c152dfc
commit
4f5bff993b
|
@ -222,6 +222,8 @@ mod comm;
|
|||
mod task {
|
||||
mod local_data;
|
||||
mod local_data_priv;
|
||||
mod spawn;
|
||||
mod rt;
|
||||
}
|
||||
mod future;
|
||||
mod pipes;
|
||||
|
|
|
@ -75,9 +75,8 @@ export ThreadPerTask;
|
|||
export ManualThreads;
|
||||
export PlatformThread;
|
||||
|
||||
macro_rules! move_it (
|
||||
{ $x:expr } => { unsafe { let y <- *ptr::addr_of($x); move y } }
|
||||
)
|
||||
use rt::task_id;
|
||||
use rt::rust_task;
|
||||
|
||||
/// A handle to a task
|
||||
enum Task {
|
||||
|
@ -498,7 +497,7 @@ impl TaskBuilder {
|
|||
mut notify_chan: move notify_chan,
|
||||
sched: x.opts.sched
|
||||
};
|
||||
spawn_raw(move opts, x.gen_body(move f));
|
||||
spawn::spawn_raw(move opts, x.gen_body(move f));
|
||||
}
|
||||
/// Runs a task, while transfering ownership of one argument to the child.
|
||||
fn spawn_with<A: Send>(+arg: A, +f: fn~(+A)) {
|
||||
|
@ -698,8 +697,8 @@ fn try<T:Send>(+f: fn~() -> T) -> Result<T,()> {
|
|||
fn yield() {
|
||||
//! Yield control to the task scheduler
|
||||
|
||||
let task_ = rustrt::rust_get_task();
|
||||
let killed = rustrt::rust_task_yield(task_);
|
||||
let task_ = rt::rust_get_task();
|
||||
let killed = rt::rust_task_yield(task_);
|
||||
if killed && !failing() {
|
||||
fail ~"killed";
|
||||
}
|
||||
|
@ -708,13 +707,13 @@ fn yield() {
|
|||
fn failing() -> bool {
|
||||
//! True if the running task has failed
|
||||
|
||||
rustrt::rust_task_is_unwinding(rustrt::rust_get_task())
|
||||
rt::rust_task_is_unwinding(rt::rust_get_task())
|
||||
}
|
||||
|
||||
fn get_task() -> Task {
|
||||
//! Get a handle to the running task
|
||||
|
||||
TaskHandle(rustrt::get_task_id())
|
||||
TaskHandle(rt::get_task_id())
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -735,7 +734,7 @@ fn get_task() -> Task {
|
|||
unsafe fn unkillable<U>(f: fn() -> U) -> U {
|
||||
struct AllowFailure {
|
||||
t: *rust_task,
|
||||
drop { rustrt::rust_task_allow_kill(self.t); }
|
||||
drop { rt::rust_task_allow_kill(self.t); }
|
||||
}
|
||||
|
||||
fn AllowFailure(t: *rust_task) -> AllowFailure{
|
||||
|
@ -744,9 +743,9 @@ unsafe fn unkillable<U>(f: fn() -> U) -> U {
|
|||
}
|
||||
}
|
||||
|
||||
let t = rustrt::rust_get_task();
|
||||
let t = rt::rust_get_task();
|
||||
let _allow_failure = AllowFailure(t);
|
||||
rustrt::rust_task_inhibit_kill(t);
|
||||
rt::rust_task_inhibit_kill(t);
|
||||
f()
|
||||
}
|
||||
|
||||
|
@ -754,7 +753,7 @@ unsafe fn unkillable<U>(f: fn() -> U) -> U {
|
|||
unsafe fn rekillable<U>(f: fn() -> U) -> U {
|
||||
struct DisallowFailure {
|
||||
t: *rust_task,
|
||||
drop { rustrt::rust_task_inhibit_kill(self.t); }
|
||||
drop { rt::rust_task_inhibit_kill(self.t); }
|
||||
}
|
||||
|
||||
fn DisallowFailure(t: *rust_task) -> DisallowFailure {
|
||||
|
@ -763,9 +762,9 @@ unsafe fn rekillable<U>(f: fn() -> U) -> U {
|
|||
}
|
||||
}
|
||||
|
||||
let t = rustrt::rust_get_task();
|
||||
let t = rt::rust_get_task();
|
||||
let _allow_failure = DisallowFailure(t);
|
||||
rustrt::rust_task_allow_kill(t);
|
||||
rt::rust_task_allow_kill(t);
|
||||
f()
|
||||
}
|
||||
|
||||
|
@ -777,8 +776,8 @@ unsafe fn atomically<U>(f: fn() -> U) -> U {
|
|||
struct DeferInterrupts {
|
||||
t: *rust_task,
|
||||
drop {
|
||||
rustrt::rust_task_allow_yield(self.t);
|
||||
rustrt::rust_task_allow_kill(self.t);
|
||||
rt::rust_task_allow_yield(self.t);
|
||||
rt::rust_task_allow_kill(self.t);
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -788,712 +787,13 @@ unsafe fn atomically<U>(f: fn() -> U) -> U {
|
|||
}
|
||||
}
|
||||
|
||||
let t = rustrt::rust_get_task();
|
||||
let t = rt::rust_get_task();
|
||||
let _interrupts = DeferInterrupts(t);
|
||||
rustrt::rust_task_inhibit_kill(t);
|
||||
rustrt::rust_task_inhibit_yield(t);
|
||||
rt::rust_task_inhibit_kill(t);
|
||||
rt::rust_task_inhibit_yield(t);
|
||||
f()
|
||||
}
|
||||
|
||||
/* **************************************************************************
|
||||
* Spawning & linked failure
|
||||
*
|
||||
* Several data structures are involved in task management to allow properly
|
||||
* propagating failure across linked/supervised tasks.
|
||||
*
|
||||
* (1) The "taskgroup_arc" is an unsafe::exclusive which contains a hashset of
|
||||
* all tasks that are part of the group. Some tasks are 'members', which
|
||||
* means if they fail, they will kill everybody else in the taskgroup.
|
||||
* Other tasks are 'descendants', which means they will not kill tasks
|
||||
* from this group, but can be killed by failing members.
|
||||
*
|
||||
* A new one of these is created each spawn_linked or spawn_supervised.
|
||||
*
|
||||
* (2) The "tcb" is a per-task control structure that tracks a task's spawn
|
||||
* configuration. It contains a reference to its taskgroup_arc, a
|
||||
* reference to its node in the ancestor list (below), a flag for
|
||||
* whether it's part of the 'main'/'root' taskgroup, and an optionally
|
||||
* configured notification port. These are stored in TLS.
|
||||
*
|
||||
* (3) The "ancestor_list" is a cons-style list of unsafe::exclusives which
|
||||
* tracks 'generations' of taskgroups -- a group's ancestors are groups
|
||||
* which (directly or transitively) spawn_supervised-ed them. Each task
|
||||
* is recorded in the 'descendants' of each of its ancestor groups.
|
||||
*
|
||||
* Spawning a supervised task is O(n) in the number of generations still
|
||||
* alive, and exiting (by success or failure) that task is also O(n).
|
||||
*
|
||||
* This diagram depicts the references between these data structures:
|
||||
*
|
||||
* linked_________________________________
|
||||
* ___/ _________ \___
|
||||
* / \ | group X | / \
|
||||
* ( A ) - - - - - - - > | {A,B} {}|< - - -( B )
|
||||
* \___/ |_________| \___/
|
||||
* unlinked
|
||||
* | __ (nil)
|
||||
* | //| The following code causes this:
|
||||
* |__ // /\ _________
|
||||
* / \ // || | group Y | fn taskA() {
|
||||
* ( C )- - - ||- - - > |{C} {D,E}| spawn(taskB);
|
||||
* \___/ / \=====> |_________| spawn_unlinked(taskC);
|
||||
* supervise /gen \ ...
|
||||
* | __ \ 00 / }
|
||||
* | //| \__/ fn taskB() { ... }
|
||||
* |__ // /\ _________ fn taskC() {
|
||||
* / \/ || | group Z | spawn_supervised(taskD);
|
||||
* ( D )- - - ||- - - > | {D} {E} | ...
|
||||
* \___/ / \=====> |_________| }
|
||||
* supervise /gen \ fn taskD() {
|
||||
* | __ \ 01 / spawn_supervised(taskE);
|
||||
* | //| \__/ ...
|
||||
* |__ // _________ }
|
||||
* / \/ | group W | fn taskE() { ... }
|
||||
* ( E )- - - - - - - > | {E} {} |
|
||||
* \___/ |_________|
|
||||
*
|
||||
* "tcb" "taskgroup_arc"
|
||||
* "ancestor_list"
|
||||
*
|
||||
****************************************************************************/
|
||||
|
||||
#[allow(non_camel_case_types)] // runtime type
|
||||
type sched_id = int;
|
||||
#[allow(non_camel_case_types)] // runtime type
|
||||
type task_id = int;
|
||||
|
||||
// These are both opaque runtime/compiler types that we don't know the
|
||||
// structure of and should only deal with via unsafe pointer
|
||||
#[allow(non_camel_case_types)] // runtime type
|
||||
type rust_task = libc::c_void;
|
||||
#[allow(non_camel_case_types)] // runtime type
|
||||
type rust_closure = libc::c_void;
|
||||
|
||||
type TaskSet = send_map::linear::LinearMap<*rust_task,()>;
|
||||
|
||||
fn new_taskset() -> TaskSet {
|
||||
send_map::linear::LinearMap()
|
||||
}
|
||||
fn taskset_insert(tasks: &mut TaskSet, task: *rust_task) {
|
||||
let didnt_overwrite = tasks.insert(task, ());
|
||||
assert didnt_overwrite;
|
||||
}
|
||||
fn taskset_remove(tasks: &mut TaskSet, task: *rust_task) {
|
||||
let was_present = tasks.remove(&task);
|
||||
assert was_present;
|
||||
}
|
||||
fn taskset_each(tasks: &TaskSet, blk: fn(+*rust_task) -> bool) {
|
||||
tasks.each_key(|k| blk(*k))
|
||||
}
|
||||
|
||||
// One of these per group of linked-failure tasks.
|
||||
type TaskGroupData = {
|
||||
// All tasks which might kill this group. When this is empty, the group
|
||||
// can be "GC"ed (i.e., its link in the ancestor list can be removed).
|
||||
mut members: TaskSet,
|
||||
// All tasks unidirectionally supervised by (directly or transitively)
|
||||
// tasks in this group.
|
||||
mut descendants: TaskSet,
|
||||
};
|
||||
type TaskGroupArc = private::Exclusive<Option<TaskGroupData>>;
|
||||
|
||||
type TaskGroupInner = &mut Option<TaskGroupData>;
|
||||
|
||||
// A taskgroup is 'dead' when nothing can cause it to fail; only members can.
|
||||
pure fn taskgroup_is_dead(tg: &TaskGroupData) -> bool {
|
||||
(&tg.members).is_empty()
|
||||
}
|
||||
|
||||
// A list-like structure by which taskgroups keep track of all ancestor groups
|
||||
// which may kill them. Needed for tasks to be able to remove themselves from
|
||||
// ancestor groups upon exit. The list has a node for each "generation", and
|
||||
// ends either at the root taskgroup (which has no ancestors) or at a
|
||||
// taskgroup which was spawned-unlinked. Tasks from intermediate generations
|
||||
// have references to the middle of the list; when intermediate generations
|
||||
// die, their node in the list will be collected at a descendant's spawn-time.
|
||||
type AncestorNode = {
|
||||
// Since the ancestor list is recursive, we end up with references to
|
||||
// exclusives within other exclusives. This is dangerous business (if
|
||||
// circular references arise, deadlock and memory leaks are imminent).
|
||||
// Hence we assert that this counter monotonically decreases as we
|
||||
// approach the tail of the list.
|
||||
// FIXME(#3068): Make the generation counter togglable with #[cfg(debug)].
|
||||
generation: uint,
|
||||
// Should really be an immutable non-option. This way appeases borrowck.
|
||||
mut parent_group: Option<TaskGroupArc>,
|
||||
// Recursive rest of the list.
|
||||
mut ancestors: AncestorList,
|
||||
};
|
||||
enum AncestorList = Option<private::Exclusive<AncestorNode>>;
|
||||
|
||||
// Accessors for taskgroup arcs and ancestor arcs that wrap the unsafety.
|
||||
#[inline(always)]
|
||||
fn access_group<U>(x: &TaskGroupArc, blk: fn(TaskGroupInner) -> U) -> U {
|
||||
unsafe { x.with(blk) }
|
||||
}
|
||||
|
||||
#[inline(always)]
|
||||
fn access_ancestors<U>(x: &private::Exclusive<AncestorNode>,
|
||||
blk: fn(x: &mut AncestorNode) -> U) -> U {
|
||||
unsafe { x.with(blk) }
|
||||
}
|
||||
|
||||
// Iterates over an ancestor list.
|
||||
// (1) Runs forward_blk on each ancestral taskgroup in the list
|
||||
// (2) If forward_blk "break"s, runs optional bail_blk on all ancestral
|
||||
// taskgroups that forward_blk already ran on successfully (Note: bail_blk
|
||||
// is NOT called on the block that forward_blk broke on!).
|
||||
// (3) As a bonus, coalesces away all 'dead' taskgroup nodes in the list.
|
||||
// FIXME(#2190): Change Option<fn@(...)> to Option<fn&(...)>, to save on
|
||||
// allocations. Once that bug is fixed, changing the sigil should suffice.
|
||||
fn each_ancestor(list: &mut AncestorList,
|
||||
bail_opt: Option<fn@(TaskGroupInner)>,
|
||||
forward_blk: fn(TaskGroupInner) -> bool)
|
||||
-> bool {
|
||||
// "Kickoff" call - there was no last generation.
|
||||
return !coalesce(list, bail_opt, forward_blk, uint::max_value);
|
||||
|
||||
// Recursively iterates, and coalesces afterwards if needed. Returns
|
||||
// whether or not unwinding is needed (i.e., !successful iteration).
|
||||
fn coalesce(list: &mut AncestorList,
|
||||
bail_opt: Option<fn@(TaskGroupInner)>,
|
||||
forward_blk: fn(TaskGroupInner) -> bool,
|
||||
last_generation: uint) -> bool {
|
||||
// Need to swap the list out to use it, to appease borrowck.
|
||||
let tmp_list = util::replace(list, AncestorList(None));
|
||||
let (coalesce_this, early_break) =
|
||||
iterate(&tmp_list, bail_opt, forward_blk, last_generation);
|
||||
// What should our next ancestor end up being?
|
||||
if coalesce_this.is_some() {
|
||||
// Needed coalesce. Our next ancestor becomes our old
|
||||
// ancestor's next ancestor. ("next = old_next->next;")
|
||||
*list <- option::unwrap(move coalesce_this);
|
||||
} else {
|
||||
// No coalesce; restore from tmp. ("next = old_next;")
|
||||
*list <- tmp_list;
|
||||
}
|
||||
return early_break;
|
||||
}
|
||||
|
||||
// Returns an optional list-to-coalesce and whether unwinding is needed.
|
||||
// Option<ancestor_list>:
|
||||
// Whether or not the ancestor taskgroup being iterated over is
|
||||
// dead or not; i.e., it has no more tasks left in it, whether or not
|
||||
// it has descendants. If dead, the caller shall coalesce it away.
|
||||
// bool:
|
||||
// True if the supplied block did 'break', here or in any recursive
|
||||
// calls. If so, must call the unwinder on all previous nodes.
|
||||
fn iterate(ancestors: &AncestorList,
|
||||
bail_opt: Option<fn@(TaskGroupInner)>,
|
||||
forward_blk: fn(TaskGroupInner) -> bool,
|
||||
last_generation: uint) -> (Option<AncestorList>, bool) {
|
||||
// At each step of iteration, three booleans are at play which govern
|
||||
// how the iteration should behave.
|
||||
// 'nobe_is_dead' - Should the list should be coalesced at this point?
|
||||
// Largely unrelated to the other two.
|
||||
// 'need_unwind' - Should we run the bail_blk at this point? (i.e.,
|
||||
// do_continue was false not here, but down the line)
|
||||
// 'do_continue' - Did the forward_blk succeed at this point? (i.e.,
|
||||
// should we recurse? or should our callers unwind?)
|
||||
|
||||
// The map defaults to None, because if ancestors is None, we're at
|
||||
// the end of the list, which doesn't make sense to coalesce.
|
||||
return do (**ancestors).map_default((None,false)) |ancestor_arc| {
|
||||
// NB: Takes a lock! (this ancestor node)
|
||||
do access_ancestors(&ancestor_arc) |nobe| {
|
||||
// Check monotonicity
|
||||
assert last_generation > nobe.generation;
|
||||
/*##########################################################*
|
||||
* Step 1: Look at this ancestor group (call iterator block).
|
||||
*##########################################################*/
|
||||
let mut nobe_is_dead = false;
|
||||
let do_continue =
|
||||
// NB: Takes a lock! (this ancestor node's parent group)
|
||||
do with_parent_tg(&mut nobe.parent_group) |tg_opt| {
|
||||
// Decide whether this group is dead. Note that the
|
||||
// group being *dead* is disjoint from it *failing*.
|
||||
nobe_is_dead = match *tg_opt {
|
||||
Some(ref tg) => taskgroup_is_dead(tg),
|
||||
None => nobe_is_dead
|
||||
};
|
||||
// Call iterator block. (If the group is dead, it's
|
||||
// safe to skip it. This will leave our *rust_task
|
||||
// hanging around in the group even after it's freed,
|
||||
// but that's ok because, by virtue of the group being
|
||||
// dead, nobody will ever kill-all (foreach) over it.)
|
||||
if nobe_is_dead { true } else { forward_blk(tg_opt) }
|
||||
};
|
||||
/*##########################################################*
|
||||
* Step 2: Recurse on the rest of the list; maybe coalescing.
|
||||
*##########################################################*/
|
||||
// 'need_unwind' is only set if blk returned true above, *and*
|
||||
// the recursive call early-broke.
|
||||
let mut need_unwind = false;
|
||||
if do_continue {
|
||||
// NB: Takes many locks! (ancestor nodes & parent groups)
|
||||
need_unwind = coalesce(&mut nobe.ancestors, bail_opt,
|
||||
forward_blk, nobe.generation);
|
||||
}
|
||||
/*##########################################################*
|
||||
* Step 3: Maybe unwind; compute return info for our caller.
|
||||
*##########################################################*/
|
||||
if need_unwind && !nobe_is_dead {
|
||||
do bail_opt.iter |bail_blk| {
|
||||
do with_parent_tg(&mut nobe.parent_group) |tg_opt| {
|
||||
bail_blk(tg_opt)
|
||||
}
|
||||
}
|
||||
}
|
||||
// Decide whether our caller should unwind.
|
||||
need_unwind = need_unwind || !do_continue;
|
||||
// Tell caller whether or not to coalesce and/or unwind
|
||||
if nobe_is_dead {
|
||||
// Swap the list out here; the caller replaces us with it.
|
||||
let rest = util::replace(&mut nobe.ancestors,
|
||||
AncestorList(None));
|
||||
(Some(move rest), need_unwind)
|
||||
} else {
|
||||
(None, need_unwind)
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
// Wrapper around exclusive::with that appeases borrowck.
|
||||
fn with_parent_tg<U>(parent_group: &mut Option<TaskGroupArc>,
|
||||
blk: fn(TaskGroupInner) -> U) -> U {
|
||||
// If this trips, more likely the problem is 'blk' failed inside.
|
||||
let tmp_arc = option::swap_unwrap(parent_group);
|
||||
let result = do access_group(&tmp_arc) |tg_opt| { blk(tg_opt) };
|
||||
*parent_group <- Some(move tmp_arc);
|
||||
move result
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// One of these per task.
|
||||
struct TCB {
|
||||
me: *rust_task,
|
||||
// List of tasks with whose fates this one's is intertwined.
|
||||
tasks: TaskGroupArc, // 'none' means the group has failed.
|
||||
// Lists of tasks who will kill us if they fail, but whom we won't kill.
|
||||
mut ancestors: AncestorList,
|
||||
is_main: bool,
|
||||
notifier: Option<AutoNotify>,
|
||||
// Runs on task exit.
|
||||
drop {
|
||||
// If we are failing, the whole taskgroup needs to die.
|
||||
if rustrt::rust_task_is_unwinding(self.me) {
|
||||
self.notifier.iter(|x| { x.failed = true; });
|
||||
// Take everybody down with us.
|
||||
do access_group(&self.tasks) |tg| {
|
||||
kill_taskgroup(tg, self.me, self.is_main);
|
||||
}
|
||||
} else {
|
||||
// Remove ourselves from the group(s).
|
||||
do access_group(&self.tasks) |tg| {
|
||||
leave_taskgroup(tg, self.me, true);
|
||||
}
|
||||
}
|
||||
// It doesn't matter whether this happens before or after dealing with
|
||||
// our own taskgroup, so long as both happen before we die. We need to
|
||||
// remove ourself from every ancestor we can, so no cleanup; no break.
|
||||
for each_ancestor(&mut self.ancestors, None) |ancestor_group| {
|
||||
leave_taskgroup(ancestor_group, self.me, false);
|
||||
};
|
||||
}
|
||||
}
|
||||
|
||||
fn TCB(me: *rust_task, +tasks: TaskGroupArc, +ancestors: AncestorList,
|
||||
is_main: bool, +notifier: Option<AutoNotify>) -> TCB {
|
||||
|
||||
let notifier = move notifier;
|
||||
notifier.iter(|x| { x.failed = false; });
|
||||
|
||||
TCB {
|
||||
me: me,
|
||||
tasks: tasks,
|
||||
ancestors: ancestors,
|
||||
is_main: is_main,
|
||||
notifier: move notifier
|
||||
}
|
||||
}
|
||||
|
||||
struct AutoNotify {
|
||||
notify_chan: Chan<Notification>,
|
||||
mut failed: bool,
|
||||
drop {
|
||||
let result = if self.failed { Failure } else { Success };
|
||||
self.notify_chan.send(Exit(get_task(), result));
|
||||
}
|
||||
}
|
||||
|
||||
fn AutoNotify(+chan: Chan<Notification>) -> AutoNotify {
|
||||
AutoNotify {
|
||||
notify_chan: chan,
|
||||
failed: true // Un-set above when taskgroup successfully made.
|
||||
}
|
||||
}
|
||||
|
||||
fn enlist_in_taskgroup(state: TaskGroupInner, me: *rust_task,
|
||||
is_member: bool) -> bool {
|
||||
let newstate = util::replace(state, None);
|
||||
// If 'None', the group was failing. Can't enlist.
|
||||
if newstate.is_some() {
|
||||
let group = option::unwrap(move newstate);
|
||||
taskset_insert(if is_member { &mut group.members }
|
||||
else { &mut group.descendants }, me);
|
||||
*state = Some(move group);
|
||||
true
|
||||
} else {
|
||||
false
|
||||
}
|
||||
}
|
||||
|
||||
// NB: Runs in destructor/post-exit context. Can't 'fail'.
|
||||
fn leave_taskgroup(state: TaskGroupInner, me: *rust_task, is_member: bool) {
|
||||
let newstate = util::replace(state, None);
|
||||
// If 'None', already failing and we've already gotten a kill signal.
|
||||
if newstate.is_some() {
|
||||
let group = option::unwrap(move newstate);
|
||||
taskset_remove(if is_member { &mut group.members }
|
||||
else { &mut group.descendants }, me);
|
||||
*state = Some(move group);
|
||||
}
|
||||
}
|
||||
|
||||
// NB: Runs in destructor/post-exit context. Can't 'fail'.
|
||||
fn kill_taskgroup(state: TaskGroupInner, me: *rust_task, is_main: bool) {
|
||||
// NB: We could do the killing iteration outside of the group arc, by
|
||||
// having "let mut newstate" here, swapping inside, and iterating after.
|
||||
// But that would let other exiting tasks fall-through and exit while we
|
||||
// were trying to kill them, causing potential use-after-free. A task's
|
||||
// presence in the arc guarantees it's alive only while we hold the lock,
|
||||
// so if we're failing, all concurrently exiting tasks must wait for us.
|
||||
// To do it differently, we'd have to use the runtime's task refcounting,
|
||||
// but that could leave task structs around long after their task exited.
|
||||
let newstate = util::replace(state, None);
|
||||
// Might already be None, if Somebody is failing simultaneously.
|
||||
// That's ok; only one task needs to do the dirty work. (Might also
|
||||
// see 'None' if Somebody already failed and we got a kill signal.)
|
||||
if newstate.is_some() {
|
||||
let group = option::unwrap(move newstate);
|
||||
for taskset_each(&group.members) |+sibling| {
|
||||
// Skip self - killing ourself won't do much good.
|
||||
if sibling != me {
|
||||
rustrt::rust_task_kill_other(sibling);
|
||||
}
|
||||
}
|
||||
for taskset_each(&group.descendants) |+child| {
|
||||
assert child != me;
|
||||
rustrt::rust_task_kill_other(child);
|
||||
}
|
||||
// Only one task should ever do this.
|
||||
if is_main {
|
||||
rustrt::rust_task_kill_all(me);
|
||||
}
|
||||
// Do NOT restore state to Some(..)! It stays None to indicate
|
||||
// that the whole taskgroup is failing, to forbid new spawns.
|
||||
}
|
||||
// (note: multiple tasks may reach this point)
|
||||
}
|
||||
|
||||
// FIXME (#2912): Work around core-vs-coretest function duplication. Can't use
|
||||
// a proper closure because the #[test]s won't understand. Have to fake it.
|
||||
macro_rules! taskgroup_key (
|
||||
// Use a "code pointer" value that will never be a real code pointer.
|
||||
() => (cast::transmute((-2 as uint, 0u)))
|
||||
)
|
||||
|
||||
fn gen_child_taskgroup(linked: bool, supervised: bool)
|
||||
-> (TaskGroupArc, AncestorList, bool) {
|
||||
let spawner = rustrt::rust_get_task();
|
||||
/*######################################################################*
|
||||
* Step 1. Get spawner's taskgroup info.
|
||||
*######################################################################*/
|
||||
let spawner_group = match unsafe { local_get(spawner,
|
||||
taskgroup_key!()) } {
|
||||
None => {
|
||||
// Main task, doing first spawn ever. Lazily initialise here.
|
||||
let mut members = new_taskset();
|
||||
taskset_insert(&mut members, spawner);
|
||||
let tasks =
|
||||
private::exclusive(Some({ mut members: move members,
|
||||
mut descendants: new_taskset() }));
|
||||
// Main task/group has no ancestors, no notifier, etc.
|
||||
let group =
|
||||
@TCB(spawner, move tasks, AncestorList(None), true, None);
|
||||
unsafe {
|
||||
local_set(spawner, taskgroup_key!(), group);
|
||||
}
|
||||
group
|
||||
}
|
||||
Some(group) => group
|
||||
};
|
||||
/*######################################################################*
|
||||
* Step 2. Process spawn options for child.
|
||||
*######################################################################*/
|
||||
return if linked {
|
||||
// Child is in the same group as spawner.
|
||||
let g = spawner_group.tasks.clone();
|
||||
// Child's ancestors are spawner's ancestors.
|
||||
let a = share_ancestors(&mut spawner_group.ancestors);
|
||||
// Propagate main-ness.
|
||||
(move g, move a, spawner_group.is_main)
|
||||
} else {
|
||||
// Child is in a separate group from spawner.
|
||||
let g = private::exclusive(Some({ mut members: new_taskset(),
|
||||
mut descendants: new_taskset() }));
|
||||
let a = if supervised {
|
||||
// Child's ancestors start with the spawner.
|
||||
let old_ancestors = share_ancestors(&mut spawner_group.ancestors);
|
||||
// FIXME(#3068) - The generation counter is only used for a debug
|
||||
// assertion, but initialising it requires locking a mutex. Hence
|
||||
// it should be enabled only in debug builds.
|
||||
let new_generation =
|
||||
match *old_ancestors {
|
||||
Some(arc) => access_ancestors(&arc, |a| a.generation+1),
|
||||
None => 0 // the actual value doesn't really matter.
|
||||
};
|
||||
assert new_generation < uint::max_value;
|
||||
// Build a new node in the ancestor list.
|
||||
AncestorList(Some(private::exclusive(
|
||||
{ generation: new_generation,
|
||||
mut parent_group: Some(spawner_group.tasks.clone()),
|
||||
mut ancestors: move old_ancestors })))
|
||||
} else {
|
||||
// Child has no ancestors.
|
||||
AncestorList(None)
|
||||
};
|
||||
(move g, move a, false)
|
||||
};
|
||||
|
||||
fn share_ancestors(ancestors: &mut AncestorList) -> AncestorList {
|
||||
// Appease the borrow-checker. Really this wants to be written as:
|
||||
// match ancestors
|
||||
// Some(ancestor_arc) { ancestor_list(Some(ancestor_arc.clone())) }
|
||||
// None { ancestor_list(None) }
|
||||
let tmp = util::replace(&mut **ancestors, None);
|
||||
if tmp.is_some() {
|
||||
let ancestor_arc = option::unwrap(move tmp);
|
||||
let result = ancestor_arc.clone();
|
||||
**ancestors <- Some(move ancestor_arc);
|
||||
AncestorList(Some(move result))
|
||||
} else {
|
||||
AncestorList(None)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn spawn_raw(+opts: TaskOpts, +f: fn~()) {
|
||||
let (child_tg, ancestors, is_main) =
|
||||
gen_child_taskgroup(opts.linked, opts.supervised);
|
||||
|
||||
unsafe {
|
||||
let child_data = ~mut Some((move child_tg, move ancestors, move f));
|
||||
// Being killed with the unsafe task/closure pointers would leak them.
|
||||
do unkillable {
|
||||
// Agh. Get move-mode items into the closure. FIXME (#2829)
|
||||
let (child_tg, ancestors, f) = option::swap_unwrap(child_data);
|
||||
// Create child task.
|
||||
let new_task = match opts.sched {
|
||||
None => rustrt::new_task(),
|
||||
Some(sched_opts) => new_task_in_new_sched(sched_opts)
|
||||
};
|
||||
assert !new_task.is_null();
|
||||
// Getting killed after here would leak the task.
|
||||
let mut notify_chan = if opts.notify_chan.is_none() {
|
||||
None
|
||||
} else {
|
||||
Some(option::swap_unwrap(&mut opts.notify_chan))
|
||||
};
|
||||
|
||||
let child_wrapper = make_child_wrapper(new_task, move child_tg,
|
||||
move ancestors, is_main, move notify_chan, move f);
|
||||
let fptr = ptr::addr_of(child_wrapper);
|
||||
let closure: *rust_closure = cast::reinterpret_cast(&fptr);
|
||||
|
||||
// Getting killed between these two calls would free the child's
|
||||
// closure. (Reordering them wouldn't help - then getting killed
|
||||
// between them would leak.)
|
||||
rustrt::start_task(new_task, closure);
|
||||
cast::forget(move child_wrapper);
|
||||
}
|
||||
}
|
||||
|
||||
// This function returns a closure-wrapper that we pass to the child task.
|
||||
// (1) It sets up the notification channel.
|
||||
// (2) It attempts to enlist in the child's group and all ancestor groups.
|
||||
// (3a) If any of those fails, it leaves all groups, and does nothing.
|
||||
// (3b) Otherwise it builds a task control structure and puts it in TLS,
|
||||
// (4) ...and runs the provided body function.
|
||||
fn make_child_wrapper(child: *rust_task, +child_arc: TaskGroupArc,
|
||||
+ancestors: AncestorList, is_main: bool,
|
||||
+notify_chan: Option<Chan<Notification>>,
|
||||
+f: fn~()) -> fn~() {
|
||||
let child_data = ~mut Some((move child_arc, move ancestors));
|
||||
return fn~(move notify_chan, move child_data, move f) {
|
||||
// Agh. Get move-mode items into the closure. FIXME (#2829)
|
||||
let mut (child_arc, ancestors) = option::swap_unwrap(child_data);
|
||||
// Child task runs this code.
|
||||
|
||||
// Even if the below code fails to kick the child off, we must
|
||||
// send Something on the notify channel.
|
||||
|
||||
//let mut notifier = None;//notify_chan.map(|c| AutoNotify(c));
|
||||
let notifier = match notify_chan {
|
||||
Some(notify_chan_value) => {
|
||||
let moved_ncv = move_it!(notify_chan_value);
|
||||
Some(AutoNotify(move moved_ncv))
|
||||
}
|
||||
_ => None
|
||||
};
|
||||
|
||||
if enlist_many(child, &child_arc, &mut ancestors) {
|
||||
let group = @TCB(child, move child_arc, move ancestors,
|
||||
is_main, move notifier);
|
||||
unsafe {
|
||||
local_set(child, taskgroup_key!(), group);
|
||||
}
|
||||
|
||||
// Run the child's body.
|
||||
f();
|
||||
|
||||
// TLS cleanup code will exit the taskgroup.
|
||||
}
|
||||
|
||||
// Run the box annihilator.
|
||||
// XXX: Crashy.
|
||||
// unsafe { cleanup::annihilate(); }
|
||||
};
|
||||
|
||||
// Set up membership in taskgroup and descendantship in all ancestor
|
||||
// groups. If any enlistment fails, Some task was already failing, so
|
||||
// don't let the child task run, and undo every successful enlistment.
|
||||
fn enlist_many(child: *rust_task, child_arc: &TaskGroupArc,
|
||||
ancestors: &mut AncestorList) -> bool {
|
||||
// Join this taskgroup.
|
||||
let mut result =
|
||||
do access_group(child_arc) |child_tg| {
|
||||
enlist_in_taskgroup(child_tg, child, true) // member
|
||||
};
|
||||
if result {
|
||||
// Unwinding function in case any ancestral enlisting fails
|
||||
let bail = |tg: TaskGroupInner| {
|
||||
leave_taskgroup(tg, child, false)
|
||||
};
|
||||
// Attempt to join every ancestor group.
|
||||
result =
|
||||
for each_ancestor(ancestors, Some(bail)) |ancestor_tg| {
|
||||
// Enlist as a descendant, not as an actual member.
|
||||
// Descendants don't kill ancestor groups on failure.
|
||||
if !enlist_in_taskgroup(ancestor_tg, child, false) {
|
||||
break;
|
||||
}
|
||||
};
|
||||
// If any ancestor group fails, need to exit this group too.
|
||||
if !result {
|
||||
do access_group(child_arc) |child_tg| {
|
||||
leave_taskgroup(child_tg, child, true); // member
|
||||
}
|
||||
}
|
||||
}
|
||||
result
|
||||
}
|
||||
}
|
||||
|
||||
fn new_task_in_new_sched(opts: SchedOpts) -> *rust_task {
|
||||
if opts.foreign_stack_size != None {
|
||||
fail ~"foreign_stack_size scheduler option unimplemented";
|
||||
}
|
||||
|
||||
let num_threads = match opts.mode {
|
||||
SingleThreaded => 1u,
|
||||
ThreadPerCore => rustrt::rust_num_threads(),
|
||||
ThreadPerTask => {
|
||||
fail ~"ThreadPerTask scheduling mode unimplemented"
|
||||
}
|
||||
ManualThreads(threads) => {
|
||||
if threads == 0u {
|
||||
fail ~"can not create a scheduler with no threads";
|
||||
}
|
||||
threads
|
||||
}
|
||||
PlatformThread => 0u /* Won't be used */
|
||||
};
|
||||
|
||||
let sched_id = if opts.mode != PlatformThread {
|
||||
rustrt::rust_new_sched(num_threads)
|
||||
} else {
|
||||
rustrt::rust_osmain_sched_id()
|
||||
};
|
||||
rustrt::rust_new_task_in_sched(sched_id)
|
||||
}
|
||||
}
|
||||
|
||||
extern mod rustrt {
|
||||
#[rust_stack]
|
||||
fn rust_task_yield(task: *rust_task) -> bool;
|
||||
|
||||
fn rust_get_sched_id() -> sched_id;
|
||||
fn rust_new_sched(num_threads: libc::uintptr_t) -> sched_id;
|
||||
fn rust_sched_threads() -> libc::size_t;
|
||||
fn rust_sched_current_nonlazy_threads() -> libc::size_t;
|
||||
fn rust_num_threads() -> libc::uintptr_t;
|
||||
|
||||
fn get_task_id() -> task_id;
|
||||
#[rust_stack]
|
||||
fn rust_get_task() -> *rust_task;
|
||||
|
||||
fn new_task() -> *rust_task;
|
||||
fn rust_new_task_in_sched(id: sched_id) -> *rust_task;
|
||||
|
||||
fn start_task(task: *rust_task, closure: *rust_closure);
|
||||
|
||||
fn rust_task_is_unwinding(task: *rust_task) -> bool;
|
||||
fn rust_osmain_sched_id() -> sched_id;
|
||||
#[rust_stack]
|
||||
fn rust_task_inhibit_kill(t: *rust_task);
|
||||
#[rust_stack]
|
||||
fn rust_task_allow_kill(t: *rust_task);
|
||||
#[rust_stack]
|
||||
fn rust_task_inhibit_yield(t: *rust_task);
|
||||
#[rust_stack]
|
||||
fn rust_task_allow_yield(t: *rust_task);
|
||||
fn rust_task_kill_other(task: *rust_task);
|
||||
fn rust_task_kill_all(task: *rust_task);
|
||||
|
||||
#[rust_stack]
|
||||
fn rust_get_task_local_data(task: *rust_task) -> *libc::c_void;
|
||||
#[rust_stack]
|
||||
fn rust_set_task_local_data(task: *rust_task, map: *libc::c_void);
|
||||
#[rust_stack]
|
||||
fn rust_task_local_data_atexit(task: *rust_task, cleanup_fn: *u8);
|
||||
}
|
||||
|
||||
|
||||
#[test]
|
||||
fn test_spawn_raw_simple() {
|
||||
let po = comm::Port();
|
||||
let ch = comm::Chan(po);
|
||||
do spawn_raw(default_task_opts()) {
|
||||
comm::send(ch, ());
|
||||
}
|
||||
comm::recv(po);
|
||||
}
|
||||
|
||||
#[test]
|
||||
#[ignore(cfg(windows))]
|
||||
fn test_spawn_raw_unsupervise() {
|
||||
let opts = {
|
||||
linked: false,
|
||||
mut notify_chan: None,
|
||||
.. default_task_opts()
|
||||
};
|
||||
do spawn_raw(opts) {
|
||||
fail;
|
||||
}
|
||||
}
|
||||
|
||||
#[test] #[should_fail] #[ignore(cfg(windows))]
|
||||
fn test_cant_dup_task_builder() {
|
||||
let b = task().unlinked();
|
||||
|
@ -1655,43 +955,6 @@ fn test_spawn_linked_sup_propagate_sibling() {
|
|||
fail;
|
||||
}
|
||||
|
||||
#[test]
|
||||
#[ignore(cfg(windows))]
|
||||
fn test_spawn_raw_notify_success() {
|
||||
let (task_ch, task_po) = pipes::stream();
|
||||
let (notify_ch, notify_po) = pipes::stream();
|
||||
|
||||
let opts = {
|
||||
notify_chan: Some(move notify_ch),
|
||||
.. default_task_opts()
|
||||
};
|
||||
do spawn_raw(opts) |move task_ch| {
|
||||
task_ch.send(get_task());
|
||||
}
|
||||
let task_ = task_po.recv();
|
||||
assert notify_po.recv() == Exit(task_, Success);
|
||||
}
|
||||
|
||||
#[test]
|
||||
#[ignore(cfg(windows))]
|
||||
fn test_spawn_raw_notify_failure() {
|
||||
// New bindings for these
|
||||
let (task_ch, task_po) = pipes::stream();
|
||||
let (notify_ch, notify_po) = pipes::stream();
|
||||
|
||||
let opts = {
|
||||
linked: false,
|
||||
notify_chan: Some(notify_ch),
|
||||
.. default_task_opts()
|
||||
};
|
||||
do spawn_raw(opts) {
|
||||
task_ch.send(get_task());
|
||||
fail;
|
||||
}
|
||||
let task_ = task_po.recv();
|
||||
assert notify_po.recv() == Exit(task_, Failure);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_run_basic() {
|
||||
let po = comm::Port();
|
||||
|
@ -1798,10 +1061,10 @@ fn test_spawn_sched() {
|
|||
let ch = comm::Chan(po);
|
||||
|
||||
fn f(i: int, ch: comm::Chan<()>) {
|
||||
let parent_sched_id = rustrt::rust_get_sched_id();
|
||||
let parent_sched_id = rt::rust_get_sched_id();
|
||||
|
||||
do spawn_sched(SingleThreaded) {
|
||||
let child_sched_id = rustrt::rust_get_sched_id();
|
||||
let child_sched_id = rt::rust_get_sched_id();
|
||||
assert parent_sched_id != child_sched_id;
|
||||
|
||||
if (i == 0) {
|
||||
|
@ -1822,9 +1085,9 @@ fn test_spawn_sched_childs_on_same_sched() {
|
|||
let ch = comm::Chan(po);
|
||||
|
||||
do spawn_sched(SingleThreaded) {
|
||||
let parent_sched_id = rustrt::rust_get_sched_id();
|
||||
let parent_sched_id = rt::rust_get_sched_id();
|
||||
do spawn {
|
||||
let child_sched_id = rustrt::rust_get_sched_id();
|
||||
let child_sched_id = rt::rust_get_sched_id();
|
||||
// This should be on the same scheduler
|
||||
assert parent_sched_id == child_sched_id;
|
||||
comm::send(ch, ());
|
||||
|
@ -1987,13 +1250,8 @@ fn test_unkillable() {
|
|||
let po = comm::Port();
|
||||
let ch = po.chan();
|
||||
|
||||
let opts = {
|
||||
let mut opts = default_task_opts();
|
||||
opts.linked = false;
|
||||
move opts
|
||||
};
|
||||
// We want to do this after failing
|
||||
do spawn_raw(opts) {
|
||||
do spawn_unlinked {
|
||||
for iter::repeat(10u) { yield() }
|
||||
ch.send(());
|
||||
}
|
||||
|
@ -2028,12 +1286,7 @@ fn test_unkillable_nested() {
|
|||
let (ch, po) = pipes::stream();
|
||||
|
||||
// We want to do this after failing
|
||||
let opts = {
|
||||
let mut opts = default_task_opts();
|
||||
opts.linked = false;
|
||||
move opts
|
||||
};
|
||||
do spawn_raw(opts) {
|
||||
do spawn_unlinked {
|
||||
for iter::repeat(10u) { yield() }
|
||||
ch.send(());
|
||||
}
|
||||
|
@ -2099,8 +1352,8 @@ fn test_sched_thread_per_core() {
|
|||
let (chan, port) = pipes::stream();
|
||||
|
||||
do spawn_sched(ThreadPerCore) {
|
||||
let cores = rustrt::rust_num_threads();
|
||||
let reported_threads = rustrt::rust_sched_threads();
|
||||
let cores = rt::rust_num_threads();
|
||||
let reported_threads = rt::rust_sched_threads();
|
||||
assert(cores as uint == reported_threads as uint);
|
||||
chan.send(());
|
||||
}
|
||||
|
@ -2113,9 +1366,9 @@ fn test_spawn_thread_on_demand() {
|
|||
let (chan, port) = pipes::stream();
|
||||
|
||||
do spawn_sched(ManualThreads(2)) {
|
||||
let max_threads = rustrt::rust_sched_threads();
|
||||
let max_threads = rt::rust_sched_threads();
|
||||
assert(max_threads as int == 2);
|
||||
let running_threads = rustrt::rust_sched_current_nonlazy_threads();
|
||||
let running_threads = rt::rust_sched_current_nonlazy_threads();
|
||||
assert(running_threads as int == 1);
|
||||
|
||||
let (chan2, port2) = pipes::stream();
|
||||
|
@ -2124,7 +1377,7 @@ fn test_spawn_thread_on_demand() {
|
|||
chan2.send(());
|
||||
}
|
||||
|
||||
let running_threads2 = rustrt::rust_sched_current_nonlazy_threads();
|
||||
let running_threads2 = rt::rust_sched_current_nonlazy_threads();
|
||||
assert(running_threads2 as int == 2);
|
||||
|
||||
port2.recv();
|
||||
|
|
|
@ -52,7 +52,7 @@ type LocalDataKey<T: Owned> = &fn(+@T);
|
|||
unsafe fn local_data_pop<T: Owned>(
|
||||
key: LocalDataKey<T>) -> Option<@T> {
|
||||
|
||||
local_pop(rustrt::rust_get_task(), key)
|
||||
local_pop(rt::rust_get_task(), key)
|
||||
}
|
||||
/**
|
||||
* Retrieve a task-local data value. It will also be kept alive in the
|
||||
|
@ -61,7 +61,7 @@ unsafe fn local_data_pop<T: Owned>(
|
|||
unsafe fn local_data_get<T: Owned>(
|
||||
key: LocalDataKey<T>) -> Option<@T> {
|
||||
|
||||
local_get(rustrt::rust_get_task(), key)
|
||||
local_get(rt::rust_get_task(), key)
|
||||
}
|
||||
/**
|
||||
* Store a value in task-local data. If this key already has a value,
|
||||
|
@ -70,7 +70,7 @@ unsafe fn local_data_get<T: Owned>(
|
|||
unsafe fn local_data_set<T: Owned>(
|
||||
key: LocalDataKey<T>, +data: @T) {
|
||||
|
||||
local_set(rustrt::rust_get_task(), key, data)
|
||||
local_set(rt::rust_get_task(), key, data)
|
||||
}
|
||||
/**
|
||||
* Modify a task-local data value. If the function returns 'None', the
|
||||
|
@ -80,7 +80,7 @@ unsafe fn local_data_modify<T: Owned>(
|
|||
key: LocalDataKey<T>,
|
||||
modify_fn: fn(Option<@T>) -> Option<@T>) {
|
||||
|
||||
local_modify(rustrt::rust_get_task(), key, modify_fn)
|
||||
local_modify(rt::rust_get_task(), key, modify_fn)
|
||||
}
|
||||
|
||||
#[test]
|
||||
|
|
|
@ -1,6 +1,7 @@
|
|||
#[doc(hidden)]; // FIXME #3538
|
||||
|
||||
use local_data::LocalDataKey;
|
||||
use rt::rust_task;
|
||||
|
||||
trait LocalData { }
|
||||
impl<T: Owned> @T: LocalData { }
|
||||
|
@ -34,13 +35,13 @@ unsafe fn get_task_local_map(task: *rust_task) -> TaskLocalMap {
|
|||
// NOTE: The map's box lives in TLS invisibly referenced once. Each time
|
||||
// we retrieve it for get/set, we make another reference, which get/set
|
||||
// drop when they finish. No "re-storing after modifying" is needed.
|
||||
let map_ptr = rustrt::rust_get_task_local_data(task);
|
||||
let map_ptr = rt::rust_get_task_local_data(task);
|
||||
if map_ptr.is_null() {
|
||||
let map: TaskLocalMap = @dvec::DVec();
|
||||
// Use reinterpret_cast -- transmute would take map away from us also.
|
||||
rustrt::rust_set_task_local_data(
|
||||
rt::rust_set_task_local_data(
|
||||
task, cast::reinterpret_cast(&map));
|
||||
rustrt::rust_task_local_data_atexit(task, cleanup_task_local_map);
|
||||
rt::rust_task_local_data_atexit(task, cleanup_task_local_map);
|
||||
// Also need to reference it an extra time to keep it for now.
|
||||
cast::bump_box_refcount(map);
|
||||
map
|
||||
|
|
|
@ -0,0 +1,59 @@
|
|||
/*!
|
||||
|
||||
The task interface to the runtime
|
||||
|
||||
*/
|
||||
|
||||
#[doc(hidden)]; // FIXME #3538
|
||||
|
||||
#[allow(non_camel_case_types)] // runtime type
|
||||
type sched_id = int;
|
||||
#[allow(non_camel_case_types)] // runtime type
|
||||
type task_id = int;
|
||||
|
||||
// These are both opaque runtime/compiler types that we don't know the
|
||||
// structure of and should only deal with via unsafe pointer
|
||||
#[allow(non_camel_case_types)] // runtime type
|
||||
type rust_task = libc::c_void;
|
||||
#[allow(non_camel_case_types)] // runtime type
|
||||
type rust_closure = libc::c_void;
|
||||
|
||||
extern {
|
||||
#[rust_stack]
|
||||
fn rust_task_yield(task: *rust_task) -> bool;
|
||||
|
||||
fn rust_get_sched_id() -> sched_id;
|
||||
fn rust_new_sched(num_threads: libc::uintptr_t) -> sched_id;
|
||||
fn rust_sched_threads() -> libc::size_t;
|
||||
fn rust_sched_current_nonlazy_threads() -> libc::size_t;
|
||||
fn rust_num_threads() -> libc::uintptr_t;
|
||||
|
||||
fn get_task_id() -> task_id;
|
||||
#[rust_stack]
|
||||
fn rust_get_task() -> *rust_task;
|
||||
|
||||
fn new_task() -> *rust_task;
|
||||
fn rust_new_task_in_sched(id: sched_id) -> *rust_task;
|
||||
|
||||
fn start_task(task: *rust_task, closure: *rust_closure);
|
||||
|
||||
fn rust_task_is_unwinding(task: *rust_task) -> bool;
|
||||
fn rust_osmain_sched_id() -> sched_id;
|
||||
#[rust_stack]
|
||||
fn rust_task_inhibit_kill(t: *rust_task);
|
||||
#[rust_stack]
|
||||
fn rust_task_allow_kill(t: *rust_task);
|
||||
#[rust_stack]
|
||||
fn rust_task_inhibit_yield(t: *rust_task);
|
||||
#[rust_stack]
|
||||
fn rust_task_allow_yield(t: *rust_task);
|
||||
fn rust_task_kill_other(task: *rust_task);
|
||||
fn rust_task_kill_all(task: *rust_task);
|
||||
|
||||
#[rust_stack]
|
||||
fn rust_get_task_local_data(task: *rust_task) -> *libc::c_void;
|
||||
#[rust_stack]
|
||||
fn rust_set_task_local_data(task: *rust_task, map: *libc::c_void);
|
||||
#[rust_stack]
|
||||
fn rust_task_local_data_atexit(task: *rust_task, cleanup_fn: *u8);
|
||||
}
|
|
@ -0,0 +1,691 @@
|
|||
/*!**************************************************************************
|
||||
* Spawning & linked failure
|
||||
*
|
||||
* Several data structures are involved in task management to allow properly
|
||||
* propagating failure across linked/supervised tasks.
|
||||
*
|
||||
* (1) The "taskgroup_arc" is an unsafe::exclusive which contains a hashset of
|
||||
* all tasks that are part of the group. Some tasks are 'members', which
|
||||
* means if they fail, they will kill everybody else in the taskgroup.
|
||||
* Other tasks are 'descendants', which means they will not kill tasks
|
||||
* from this group, but can be killed by failing members.
|
||||
*
|
||||
* A new one of these is created each spawn_linked or spawn_supervised.
|
||||
*
|
||||
* (2) The "tcb" is a per-task control structure that tracks a task's spawn
|
||||
* configuration. It contains a reference to its taskgroup_arc, a
|
||||
* reference to its node in the ancestor list (below), a flag for
|
||||
* whether it's part of the 'main'/'root' taskgroup, and an optionally
|
||||
* configured notification port. These are stored in TLS.
|
||||
*
|
||||
* (3) The "ancestor_list" is a cons-style list of unsafe::exclusives which
|
||||
* tracks 'generations' of taskgroups -- a group's ancestors are groups
|
||||
* which (directly or transitively) spawn_supervised-ed them. Each task
|
||||
* is recorded in the 'descendants' of each of its ancestor groups.
|
||||
*
|
||||
* Spawning a supervised task is O(n) in the number of generations still
|
||||
* alive, and exiting (by success or failure) that task is also O(n).
|
||||
*
|
||||
* This diagram depicts the references between these data structures:
|
||||
*
|
||||
* linked_________________________________
|
||||
* ___/ _________ \___
|
||||
* / \ | group X | / \
|
||||
* ( A ) - - - - - - - > | {A,B} {}|< - - -( B )
|
||||
* \___/ |_________| \___/
|
||||
* unlinked
|
||||
* | __ (nil)
|
||||
* | //| The following code causes this:
|
||||
* |__ // /\ _________
|
||||
* / \ // || | group Y | fn taskA() {
|
||||
* ( C )- - - ||- - - > |{C} {D,E}| spawn(taskB);
|
||||
* \___/ / \=====> |_________| spawn_unlinked(taskC);
|
||||
* supervise /gen \ ...
|
||||
* | __ \ 00 / }
|
||||
* | //| \__/ fn taskB() { ... }
|
||||
* |__ // /\ _________ fn taskC() {
|
||||
* / \/ || | group Z | spawn_supervised(taskD);
|
||||
* ( D )- - - ||- - - > | {D} {E} | ...
|
||||
* \___/ / \=====> |_________| }
|
||||
* supervise /gen \ fn taskD() {
|
||||
* | __ \ 01 / spawn_supervised(taskE);
|
||||
* | //| \__/ ...
|
||||
* |__ // _________ }
|
||||
* / \/ | group W | fn taskE() { ... }
|
||||
* ( E )- - - - - - - > | {E} {} |
|
||||
* \___/ |_________|
|
||||
*
|
||||
* "tcb" "taskgroup_arc"
|
||||
* "ancestor_list"
|
||||
*
|
||||
****************************************************************************/
|
||||
|
||||
#[doc(hidden)]; // FIXME #3538
|
||||
|
||||
use rt::rust_task;
|
||||
use rt::rust_closure;
|
||||
|
||||
macro_rules! move_it (
|
||||
{ $x:expr } => { unsafe { let y <- *ptr::addr_of($x); move y } }
|
||||
)
|
||||
|
||||
type TaskSet = send_map::linear::LinearMap<*rust_task,()>;
|
||||
|
||||
fn new_taskset() -> TaskSet {
|
||||
send_map::linear::LinearMap()
|
||||
}
|
||||
fn taskset_insert(tasks: &mut TaskSet, task: *rust_task) {
|
||||
let didnt_overwrite = tasks.insert(task, ());
|
||||
assert didnt_overwrite;
|
||||
}
|
||||
fn taskset_remove(tasks: &mut TaskSet, task: *rust_task) {
|
||||
let was_present = tasks.remove(&task);
|
||||
assert was_present;
|
||||
}
|
||||
fn taskset_each(tasks: &TaskSet, blk: fn(+*rust_task) -> bool) {
|
||||
tasks.each_key(|k| blk(*k))
|
||||
}
|
||||
|
||||
// One of these per group of linked-failure tasks.
|
||||
type TaskGroupData = {
|
||||
// All tasks which might kill this group. When this is empty, the group
|
||||
// can be "GC"ed (i.e., its link in the ancestor list can be removed).
|
||||
mut members: TaskSet,
|
||||
// All tasks unidirectionally supervised by (directly or transitively)
|
||||
// tasks in this group.
|
||||
mut descendants: TaskSet,
|
||||
};
|
||||
type TaskGroupArc = private::Exclusive<Option<TaskGroupData>>;
|
||||
|
||||
type TaskGroupInner = &mut Option<TaskGroupData>;
|
||||
|
||||
// A taskgroup is 'dead' when nothing can cause it to fail; only members can.
|
||||
pure fn taskgroup_is_dead(tg: &TaskGroupData) -> bool {
|
||||
(&tg.members).is_empty()
|
||||
}
|
||||
|
||||
// A list-like structure by which taskgroups keep track of all ancestor groups
|
||||
// which may kill them. Needed for tasks to be able to remove themselves from
|
||||
// ancestor groups upon exit. The list has a node for each "generation", and
|
||||
// ends either at the root taskgroup (which has no ancestors) or at a
|
||||
// taskgroup which was spawned-unlinked. Tasks from intermediate generations
|
||||
// have references to the middle of the list; when intermediate generations
|
||||
// die, their node in the list will be collected at a descendant's spawn-time.
|
||||
type AncestorNode = {
|
||||
// Since the ancestor list is recursive, we end up with references to
|
||||
// exclusives within other exclusives. This is dangerous business (if
|
||||
// circular references arise, deadlock and memory leaks are imminent).
|
||||
// Hence we assert that this counter monotonically decreases as we
|
||||
// approach the tail of the list.
|
||||
// FIXME(#3068): Make the generation counter togglable with #[cfg(debug)].
|
||||
generation: uint,
|
||||
// Should really be an immutable non-option. This way appeases borrowck.
|
||||
mut parent_group: Option<TaskGroupArc>,
|
||||
// Recursive rest of the list.
|
||||
mut ancestors: AncestorList,
|
||||
};
|
||||
enum AncestorList = Option<private::Exclusive<AncestorNode>>;
|
||||
|
||||
// Accessors for taskgroup arcs and ancestor arcs that wrap the unsafety.
|
||||
#[inline(always)]
|
||||
fn access_group<U>(x: &TaskGroupArc, blk: fn(TaskGroupInner) -> U) -> U {
|
||||
unsafe { x.with(blk) }
|
||||
}
|
||||
|
||||
#[inline(always)]
|
||||
fn access_ancestors<U>(x: &private::Exclusive<AncestorNode>,
|
||||
blk: fn(x: &mut AncestorNode) -> U) -> U {
|
||||
unsafe { x.with(blk) }
|
||||
}
|
||||
|
||||
// Iterates over an ancestor list.
|
||||
// (1) Runs forward_blk on each ancestral taskgroup in the list
|
||||
// (2) If forward_blk "break"s, runs optional bail_blk on all ancestral
|
||||
// taskgroups that forward_blk already ran on successfully (Note: bail_blk
|
||||
// is NOT called on the block that forward_blk broke on!).
|
||||
// (3) As a bonus, coalesces away all 'dead' taskgroup nodes in the list.
|
||||
// FIXME(#2190): Change Option<fn@(...)> to Option<fn&(...)>, to save on
|
||||
// allocations. Once that bug is fixed, changing the sigil should suffice.
|
||||
fn each_ancestor(list: &mut AncestorList,
|
||||
bail_opt: Option<fn@(TaskGroupInner)>,
|
||||
forward_blk: fn(TaskGroupInner) -> bool)
|
||||
-> bool {
|
||||
// "Kickoff" call - there was no last generation.
|
||||
return !coalesce(list, bail_opt, forward_blk, uint::max_value);
|
||||
|
||||
// Recursively iterates, and coalesces afterwards if needed. Returns
|
||||
// whether or not unwinding is needed (i.e., !successful iteration).
|
||||
fn coalesce(list: &mut AncestorList,
|
||||
bail_opt: Option<fn@(TaskGroupInner)>,
|
||||
forward_blk: fn(TaskGroupInner) -> bool,
|
||||
last_generation: uint) -> bool {
|
||||
// Need to swap the list out to use it, to appease borrowck.
|
||||
let tmp_list = util::replace(list, AncestorList(None));
|
||||
let (coalesce_this, early_break) =
|
||||
iterate(&tmp_list, bail_opt, forward_blk, last_generation);
|
||||
// What should our next ancestor end up being?
|
||||
if coalesce_this.is_some() {
|
||||
// Needed coalesce. Our next ancestor becomes our old
|
||||
// ancestor's next ancestor. ("next = old_next->next;")
|
||||
*list <- option::unwrap(move coalesce_this);
|
||||
} else {
|
||||
// No coalesce; restore from tmp. ("next = old_next;")
|
||||
*list <- tmp_list;
|
||||
}
|
||||
return early_break;
|
||||
}
|
||||
|
||||
// Returns an optional list-to-coalesce and whether unwinding is needed.
|
||||
// Option<ancestor_list>:
|
||||
// Whether or not the ancestor taskgroup being iterated over is
|
||||
// dead or not; i.e., it has no more tasks left in it, whether or not
|
||||
// it has descendants. If dead, the caller shall coalesce it away.
|
||||
// bool:
|
||||
// True if the supplied block did 'break', here or in any recursive
|
||||
// calls. If so, must call the unwinder on all previous nodes.
|
||||
fn iterate(ancestors: &AncestorList,
|
||||
bail_opt: Option<fn@(TaskGroupInner)>,
|
||||
forward_blk: fn(TaskGroupInner) -> bool,
|
||||
last_generation: uint) -> (Option<AncestorList>, bool) {
|
||||
// At each step of iteration, three booleans are at play which govern
|
||||
// how the iteration should behave.
|
||||
// 'nobe_is_dead' - Should the list should be coalesced at this point?
|
||||
// Largely unrelated to the other two.
|
||||
// 'need_unwind' - Should we run the bail_blk at this point? (i.e.,
|
||||
// do_continue was false not here, but down the line)
|
||||
// 'do_continue' - Did the forward_blk succeed at this point? (i.e.,
|
||||
// should we recurse? or should our callers unwind?)
|
||||
|
||||
// The map defaults to None, because if ancestors is None, we're at
|
||||
// the end of the list, which doesn't make sense to coalesce.
|
||||
return do (**ancestors).map_default((None,false)) |ancestor_arc| {
|
||||
// NB: Takes a lock! (this ancestor node)
|
||||
do access_ancestors(&ancestor_arc) |nobe| {
|
||||
// Check monotonicity
|
||||
assert last_generation > nobe.generation;
|
||||
/*##########################################################*
|
||||
* Step 1: Look at this ancestor group (call iterator block).
|
||||
*##########################################################*/
|
||||
let mut nobe_is_dead = false;
|
||||
let do_continue =
|
||||
// NB: Takes a lock! (this ancestor node's parent group)
|
||||
do with_parent_tg(&mut nobe.parent_group) |tg_opt| {
|
||||
// Decide whether this group is dead. Note that the
|
||||
// group being *dead* is disjoint from it *failing*.
|
||||
nobe_is_dead = match *tg_opt {
|
||||
Some(ref tg) => taskgroup_is_dead(tg),
|
||||
None => nobe_is_dead
|
||||
};
|
||||
// Call iterator block. (If the group is dead, it's
|
||||
// safe to skip it. This will leave our *rust_task
|
||||
// hanging around in the group even after it's freed,
|
||||
// but that's ok because, by virtue of the group being
|
||||
// dead, nobody will ever kill-all (foreach) over it.)
|
||||
if nobe_is_dead { true } else { forward_blk(tg_opt) }
|
||||
};
|
||||
/*##########################################################*
|
||||
* Step 2: Recurse on the rest of the list; maybe coalescing.
|
||||
*##########################################################*/
|
||||
// 'need_unwind' is only set if blk returned true above, *and*
|
||||
// the recursive call early-broke.
|
||||
let mut need_unwind = false;
|
||||
if do_continue {
|
||||
// NB: Takes many locks! (ancestor nodes & parent groups)
|
||||
need_unwind = coalesce(&mut nobe.ancestors, bail_opt,
|
||||
forward_blk, nobe.generation);
|
||||
}
|
||||
/*##########################################################*
|
||||
* Step 3: Maybe unwind; compute return info for our caller.
|
||||
*##########################################################*/
|
||||
if need_unwind && !nobe_is_dead {
|
||||
do bail_opt.iter |bail_blk| {
|
||||
do with_parent_tg(&mut nobe.parent_group) |tg_opt| {
|
||||
bail_blk(tg_opt)
|
||||
}
|
||||
}
|
||||
}
|
||||
// Decide whether our caller should unwind.
|
||||
need_unwind = need_unwind || !do_continue;
|
||||
// Tell caller whether or not to coalesce and/or unwind
|
||||
if nobe_is_dead {
|
||||
// Swap the list out here; the caller replaces us with it.
|
||||
let rest = util::replace(&mut nobe.ancestors,
|
||||
AncestorList(None));
|
||||
(Some(move rest), need_unwind)
|
||||
} else {
|
||||
(None, need_unwind)
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
// Wrapper around exclusive::with that appeases borrowck.
|
||||
fn with_parent_tg<U>(parent_group: &mut Option<TaskGroupArc>,
|
||||
blk: fn(TaskGroupInner) -> U) -> U {
|
||||
// If this trips, more likely the problem is 'blk' failed inside.
|
||||
let tmp_arc = option::swap_unwrap(parent_group);
|
||||
let result = do access_group(&tmp_arc) |tg_opt| { blk(tg_opt) };
|
||||
*parent_group <- Some(move tmp_arc);
|
||||
move result
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// One of these per task.
|
||||
struct TCB {
|
||||
me: *rust_task,
|
||||
// List of tasks with whose fates this one's is intertwined.
|
||||
tasks: TaskGroupArc, // 'none' means the group has failed.
|
||||
// Lists of tasks who will kill us if they fail, but whom we won't kill.
|
||||
mut ancestors: AncestorList,
|
||||
is_main: bool,
|
||||
notifier: Option<AutoNotify>,
|
||||
// Runs on task exit.
|
||||
drop {
|
||||
// If we are failing, the whole taskgroup needs to die.
|
||||
if rt::rust_task_is_unwinding(self.me) {
|
||||
self.notifier.iter(|x| { x.failed = true; });
|
||||
// Take everybody down with us.
|
||||
do access_group(&self.tasks) |tg| {
|
||||
kill_taskgroup(tg, self.me, self.is_main);
|
||||
}
|
||||
} else {
|
||||
// Remove ourselves from the group(s).
|
||||
do access_group(&self.tasks) |tg| {
|
||||
leave_taskgroup(tg, self.me, true);
|
||||
}
|
||||
}
|
||||
// It doesn't matter whether this happens before or after dealing with
|
||||
// our own taskgroup, so long as both happen before we die. We need to
|
||||
// remove ourself from every ancestor we can, so no cleanup; no break.
|
||||
for each_ancestor(&mut self.ancestors, None) |ancestor_group| {
|
||||
leave_taskgroup(ancestor_group, self.me, false);
|
||||
};
|
||||
}
|
||||
}
|
||||
|
||||
fn TCB(me: *rust_task, +tasks: TaskGroupArc, +ancestors: AncestorList,
|
||||
is_main: bool, +notifier: Option<AutoNotify>) -> TCB {
|
||||
|
||||
let notifier = move notifier;
|
||||
notifier.iter(|x| { x.failed = false; });
|
||||
|
||||
TCB {
|
||||
me: me,
|
||||
tasks: tasks,
|
||||
ancestors: ancestors,
|
||||
is_main: is_main,
|
||||
notifier: move notifier
|
||||
}
|
||||
}
|
||||
|
||||
struct AutoNotify {
|
||||
notify_chan: Chan<Notification>,
|
||||
mut failed: bool,
|
||||
drop {
|
||||
let result = if self.failed { Failure } else { Success };
|
||||
self.notify_chan.send(Exit(get_task(), result));
|
||||
}
|
||||
}
|
||||
|
||||
fn AutoNotify(+chan: Chan<Notification>) -> AutoNotify {
|
||||
AutoNotify {
|
||||
notify_chan: chan,
|
||||
failed: true // Un-set above when taskgroup successfully made.
|
||||
}
|
||||
}
|
||||
|
||||
fn enlist_in_taskgroup(state: TaskGroupInner, me: *rust_task,
|
||||
is_member: bool) -> bool {
|
||||
let newstate = util::replace(state, None);
|
||||
// If 'None', the group was failing. Can't enlist.
|
||||
if newstate.is_some() {
|
||||
let group = option::unwrap(move newstate);
|
||||
taskset_insert(if is_member { &mut group.members }
|
||||
else { &mut group.descendants }, me);
|
||||
*state = Some(move group);
|
||||
true
|
||||
} else {
|
||||
false
|
||||
}
|
||||
}
|
||||
|
||||
// NB: Runs in destructor/post-exit context. Can't 'fail'.
|
||||
fn leave_taskgroup(state: TaskGroupInner, me: *rust_task, is_member: bool) {
|
||||
let newstate = util::replace(state, None);
|
||||
// If 'None', already failing and we've already gotten a kill signal.
|
||||
if newstate.is_some() {
|
||||
let group = option::unwrap(move newstate);
|
||||
taskset_remove(if is_member { &mut group.members }
|
||||
else { &mut group.descendants }, me);
|
||||
*state = Some(move group);
|
||||
}
|
||||
}
|
||||
|
||||
// NB: Runs in destructor/post-exit context. Can't 'fail'.
|
||||
fn kill_taskgroup(state: TaskGroupInner, me: *rust_task, is_main: bool) {
|
||||
// NB: We could do the killing iteration outside of the group arc, by
|
||||
// having "let mut newstate" here, swapping inside, and iterating after.
|
||||
// But that would let other exiting tasks fall-through and exit while we
|
||||
// were trying to kill them, causing potential use-after-free. A task's
|
||||
// presence in the arc guarantees it's alive only while we hold the lock,
|
||||
// so if we're failing, all concurrently exiting tasks must wait for us.
|
||||
// To do it differently, we'd have to use the runtime's task refcounting,
|
||||
// but that could leave task structs around long after their task exited.
|
||||
let newstate = util::replace(state, None);
|
||||
// Might already be None, if Somebody is failing simultaneously.
|
||||
// That's ok; only one task needs to do the dirty work. (Might also
|
||||
// see 'None' if Somebody already failed and we got a kill signal.)
|
||||
if newstate.is_some() {
|
||||
let group = option::unwrap(move newstate);
|
||||
for taskset_each(&group.members) |+sibling| {
|
||||
// Skip self - killing ourself won't do much good.
|
||||
if sibling != me {
|
||||
rt::rust_task_kill_other(sibling);
|
||||
}
|
||||
}
|
||||
for taskset_each(&group.descendants) |+child| {
|
||||
assert child != me;
|
||||
rt::rust_task_kill_other(child);
|
||||
}
|
||||
// Only one task should ever do this.
|
||||
if is_main {
|
||||
rt::rust_task_kill_all(me);
|
||||
}
|
||||
// Do NOT restore state to Some(..)! It stays None to indicate
|
||||
// that the whole taskgroup is failing, to forbid new spawns.
|
||||
}
|
||||
// (note: multiple tasks may reach this point)
|
||||
}
|
||||
|
||||
// FIXME (#2912): Work around core-vs-coretest function duplication. Can't use
|
||||
// a proper closure because the #[test]s won't understand. Have to fake it.
|
||||
macro_rules! taskgroup_key (
|
||||
// Use a "code pointer" value that will never be a real code pointer.
|
||||
() => (cast::transmute((-2 as uint, 0u)))
|
||||
)
|
||||
|
||||
fn gen_child_taskgroup(linked: bool, supervised: bool)
|
||||
-> (TaskGroupArc, AncestorList, bool) {
|
||||
let spawner = rt::rust_get_task();
|
||||
/*######################################################################*
|
||||
* Step 1. Get spawner's taskgroup info.
|
||||
*######################################################################*/
|
||||
let spawner_group = match unsafe { local_get(spawner,
|
||||
taskgroup_key!()) } {
|
||||
None => {
|
||||
// Main task, doing first spawn ever. Lazily initialise here.
|
||||
let mut members = new_taskset();
|
||||
taskset_insert(&mut members, spawner);
|
||||
let tasks =
|
||||
private::exclusive(Some({ mut members: move members,
|
||||
mut descendants: new_taskset() }));
|
||||
// Main task/group has no ancestors, no notifier, etc.
|
||||
let group =
|
||||
@TCB(spawner, move tasks, AncestorList(None), true, None);
|
||||
unsafe {
|
||||
local_set(spawner, taskgroup_key!(), group);
|
||||
}
|
||||
group
|
||||
}
|
||||
Some(group) => group
|
||||
};
|
||||
/*######################################################################*
|
||||
* Step 2. Process spawn options for child.
|
||||
*######################################################################*/
|
||||
return if linked {
|
||||
// Child is in the same group as spawner.
|
||||
let g = spawner_group.tasks.clone();
|
||||
// Child's ancestors are spawner's ancestors.
|
||||
let a = share_ancestors(&mut spawner_group.ancestors);
|
||||
// Propagate main-ness.
|
||||
(move g, move a, spawner_group.is_main)
|
||||
} else {
|
||||
// Child is in a separate group from spawner.
|
||||
let g = private::exclusive(Some({ mut members: new_taskset(),
|
||||
mut descendants: new_taskset() }));
|
||||
let a = if supervised {
|
||||
// Child's ancestors start with the spawner.
|
||||
let old_ancestors = share_ancestors(&mut spawner_group.ancestors);
|
||||
// FIXME(#3068) - The generation counter is only used for a debug
|
||||
// assertion, but initialising it requires locking a mutex. Hence
|
||||
// it should be enabled only in debug builds.
|
||||
let new_generation =
|
||||
match *old_ancestors {
|
||||
Some(arc) => access_ancestors(&arc, |a| a.generation+1),
|
||||
None => 0 // the actual value doesn't really matter.
|
||||
};
|
||||
assert new_generation < uint::max_value;
|
||||
// Build a new node in the ancestor list.
|
||||
AncestorList(Some(private::exclusive(
|
||||
{ generation: new_generation,
|
||||
mut parent_group: Some(spawner_group.tasks.clone()),
|
||||
mut ancestors: move old_ancestors })))
|
||||
} else {
|
||||
// Child has no ancestors.
|
||||
AncestorList(None)
|
||||
};
|
||||
(move g, move a, false)
|
||||
};
|
||||
|
||||
fn share_ancestors(ancestors: &mut AncestorList) -> AncestorList {
|
||||
// Appease the borrow-checker. Really this wants to be written as:
|
||||
// match ancestors
|
||||
// Some(ancestor_arc) { ancestor_list(Some(ancestor_arc.clone())) }
|
||||
// None { ancestor_list(None) }
|
||||
let tmp = util::replace(&mut **ancestors, None);
|
||||
if tmp.is_some() {
|
||||
let ancestor_arc = option::unwrap(move tmp);
|
||||
let result = ancestor_arc.clone();
|
||||
**ancestors <- Some(move ancestor_arc);
|
||||
AncestorList(Some(move result))
|
||||
} else {
|
||||
AncestorList(None)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn spawn_raw(+opts: TaskOpts, +f: fn~()) {
|
||||
let (child_tg, ancestors, is_main) =
|
||||
gen_child_taskgroup(opts.linked, opts.supervised);
|
||||
|
||||
unsafe {
|
||||
let child_data = ~mut Some((move child_tg, move ancestors, move f));
|
||||
// Being killed with the unsafe task/closure pointers would leak them.
|
||||
do unkillable {
|
||||
// Agh. Get move-mode items into the closure. FIXME (#2829)
|
||||
let (child_tg, ancestors, f) = option::swap_unwrap(child_data);
|
||||
// Create child task.
|
||||
let new_task = match opts.sched {
|
||||
None => rt::new_task(),
|
||||
Some(sched_opts) => new_task_in_new_sched(sched_opts)
|
||||
};
|
||||
assert !new_task.is_null();
|
||||
// Getting killed after here would leak the task.
|
||||
let mut notify_chan = if opts.notify_chan.is_none() {
|
||||
None
|
||||
} else {
|
||||
Some(option::swap_unwrap(&mut opts.notify_chan))
|
||||
};
|
||||
|
||||
let child_wrapper = make_child_wrapper(new_task, move child_tg,
|
||||
move ancestors, is_main, move notify_chan, move f);
|
||||
let fptr = ptr::addr_of(child_wrapper);
|
||||
let closure: *rust_closure = cast::reinterpret_cast(&fptr);
|
||||
|
||||
// Getting killed between these two calls would free the child's
|
||||
// closure. (Reordering them wouldn't help - then getting killed
|
||||
// between them would leak.)
|
||||
rt::start_task(new_task, closure);
|
||||
cast::forget(move child_wrapper);
|
||||
}
|
||||
}
|
||||
|
||||
// This function returns a closure-wrapper that we pass to the child task.
|
||||
// (1) It sets up the notification channel.
|
||||
// (2) It attempts to enlist in the child's group and all ancestor groups.
|
||||
// (3a) If any of those fails, it leaves all groups, and does nothing.
|
||||
// (3b) Otherwise it builds a task control structure and puts it in TLS,
|
||||
// (4) ...and runs the provided body function.
|
||||
fn make_child_wrapper(child: *rust_task, +child_arc: TaskGroupArc,
|
||||
+ancestors: AncestorList, is_main: bool,
|
||||
+notify_chan: Option<Chan<Notification>>,
|
||||
+f: fn~()) -> fn~() {
|
||||
let child_data = ~mut Some((move child_arc, move ancestors));
|
||||
return fn~(move notify_chan, move child_data, move f) {
|
||||
// Agh. Get move-mode items into the closure. FIXME (#2829)
|
||||
let mut (child_arc, ancestors) = option::swap_unwrap(child_data);
|
||||
// Child task runs this code.
|
||||
|
||||
// Even if the below code fails to kick the child off, we must
|
||||
// send Something on the notify channel.
|
||||
|
||||
//let mut notifier = None;//notify_chan.map(|c| AutoNotify(c));
|
||||
let notifier = match notify_chan {
|
||||
Some(notify_chan_value) => {
|
||||
let moved_ncv = move_it!(notify_chan_value);
|
||||
Some(AutoNotify(move moved_ncv))
|
||||
}
|
||||
_ => None
|
||||
};
|
||||
|
||||
if enlist_many(child, &child_arc, &mut ancestors) {
|
||||
let group = @TCB(child, move child_arc, move ancestors,
|
||||
is_main, move notifier);
|
||||
unsafe {
|
||||
local_set(child, taskgroup_key!(), group);
|
||||
}
|
||||
|
||||
// Run the child's body.
|
||||
f();
|
||||
|
||||
// TLS cleanup code will exit the taskgroup.
|
||||
}
|
||||
|
||||
// Run the box annihilator.
|
||||
// XXX: Crashy.
|
||||
// unsafe { cleanup::annihilate(); }
|
||||
};
|
||||
|
||||
// Set up membership in taskgroup and descendantship in all ancestor
|
||||
// groups. If any enlistment fails, Some task was already failing, so
|
||||
// don't let the child task run, and undo every successful enlistment.
|
||||
fn enlist_many(child: *rust_task, child_arc: &TaskGroupArc,
|
||||
ancestors: &mut AncestorList) -> bool {
|
||||
// Join this taskgroup.
|
||||
let mut result =
|
||||
do access_group(child_arc) |child_tg| {
|
||||
enlist_in_taskgroup(child_tg, child, true) // member
|
||||
};
|
||||
if result {
|
||||
// Unwinding function in case any ancestral enlisting fails
|
||||
let bail = |tg: TaskGroupInner| {
|
||||
leave_taskgroup(tg, child, false)
|
||||
};
|
||||
// Attempt to join every ancestor group.
|
||||
result =
|
||||
for each_ancestor(ancestors, Some(bail)) |ancestor_tg| {
|
||||
// Enlist as a descendant, not as an actual member.
|
||||
// Descendants don't kill ancestor groups on failure.
|
||||
if !enlist_in_taskgroup(ancestor_tg, child, false) {
|
||||
break;
|
||||
}
|
||||
};
|
||||
// If any ancestor group fails, need to exit this group too.
|
||||
if !result {
|
||||
do access_group(child_arc) |child_tg| {
|
||||
leave_taskgroup(child_tg, child, true); // member
|
||||
}
|
||||
}
|
||||
}
|
||||
result
|
||||
}
|
||||
}
|
||||
|
||||
fn new_task_in_new_sched(opts: SchedOpts) -> *rust_task {
|
||||
if opts.foreign_stack_size != None {
|
||||
fail ~"foreign_stack_size scheduler option unimplemented";
|
||||
}
|
||||
|
||||
let num_threads = match opts.mode {
|
||||
SingleThreaded => 1u,
|
||||
ThreadPerCore => rt::rust_num_threads(),
|
||||
ThreadPerTask => {
|
||||
fail ~"ThreadPerTask scheduling mode unimplemented"
|
||||
}
|
||||
ManualThreads(threads) => {
|
||||
if threads == 0u {
|
||||
fail ~"can not create a scheduler with no threads";
|
||||
}
|
||||
threads
|
||||
}
|
||||
PlatformThread => 0u /* Won't be used */
|
||||
};
|
||||
|
||||
let sched_id = if opts.mode != PlatformThread {
|
||||
rt::rust_new_sched(num_threads)
|
||||
} else {
|
||||
rt::rust_osmain_sched_id()
|
||||
};
|
||||
rt::rust_new_task_in_sched(sched_id)
|
||||
}
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_spawn_raw_simple() {
|
||||
let po = comm::Port();
|
||||
let ch = comm::Chan(po);
|
||||
do spawn_raw(default_task_opts()) {
|
||||
comm::send(ch, ());
|
||||
}
|
||||
comm::recv(po);
|
||||
}
|
||||
|
||||
#[test]
|
||||
#[ignore(cfg(windows))]
|
||||
fn test_spawn_raw_unsupervise() {
|
||||
let opts = {
|
||||
linked: false,
|
||||
mut notify_chan: None,
|
||||
.. default_task_opts()
|
||||
};
|
||||
do spawn_raw(opts) {
|
||||
fail;
|
||||
}
|
||||
}
|
||||
|
||||
#[test]
|
||||
#[ignore(cfg(windows))]
|
||||
fn test_spawn_raw_notify_success() {
|
||||
let (task_ch, task_po) = pipes::stream();
|
||||
let (notify_ch, notify_po) = pipes::stream();
|
||||
|
||||
let opts = {
|
||||
notify_chan: Some(move notify_ch),
|
||||
.. default_task_opts()
|
||||
};
|
||||
do spawn_raw(opts) |move task_ch| {
|
||||
task_ch.send(get_task());
|
||||
}
|
||||
let task_ = task_po.recv();
|
||||
assert notify_po.recv() == Exit(task_, Success);
|
||||
}
|
||||
|
||||
#[test]
|
||||
#[ignore(cfg(windows))]
|
||||
fn test_spawn_raw_notify_failure() {
|
||||
// New bindings for these
|
||||
let (task_ch, task_po) = pipes::stream();
|
||||
let (notify_ch, notify_po) = pipes::stream();
|
||||
|
||||
let opts = {
|
||||
linked: false,
|
||||
notify_chan: Some(notify_ch),
|
||||
.. default_task_opts()
|
||||
};
|
||||
do spawn_raw(opts) {
|
||||
task_ch.send(get_task());
|
||||
fail;
|
||||
}
|
||||
let task_ = task_po.recv();
|
||||
assert notify_po.recv() == Exit(task_, Failure);
|
||||
}
|
Loading…
Reference in New Issue