use field init shorthand EVERYWHERE

Like #43008 (f668999), but _much more aggressive_.
This commit is contained in:
Zack M. Davis 2017-08-06 22:54:09 -07:00
parent 82be83cf74
commit 1b6c9605e4
281 changed files with 1376 additions and 1376 deletions

View File

@ -193,7 +193,7 @@ pub struct ShouldRun<'a> {
impl<'a> ShouldRun<'a> {
fn new(builder: &'a Builder) -> ShouldRun<'a> {
ShouldRun {
builder: builder,
builder,
paths: BTreeSet::new(),
is_really_default: true, // by default no additional conditions
}
@ -278,9 +278,9 @@ impl<'a> Builder<'a> {
};
let builder = Builder {
build: build,
build,
top_stage: build.config.stage.unwrap_or(2),
kind: kind,
kind,
cache: Cache::new(),
stack: RefCell::new(Vec::new()),
};
@ -309,9 +309,9 @@ impl<'a> Builder<'a> {
};
let builder = Builder {
build: build,
build,
top_stage: build.config.stage.unwrap_or(2),
kind: kind,
kind,
cache: Cache::new(),
stack: RefCell::new(Vec::new()),
};

View File

@ -872,7 +872,7 @@ impl Step for CrateLibrustc {
builder.ensure(CrateLibrustc {
compiler,
target: run.target,
test_kind: test_kind,
test_kind,
krate: name,
});
};
@ -934,8 +934,8 @@ impl Step for Crate {
builder.ensure(Crate {
compiler,
target: run.target,
mode: mode,
test_kind: test_kind,
mode,
test_kind,
krate: name,
});
};

View File

@ -74,13 +74,13 @@ impl Step for Std {
let from = builder.compiler(1, build.build);
builder.ensure(Std {
compiler: from,
target: target,
target,
});
println!("Uplifting stage1 std ({} -> {})", from.host, target);
builder.ensure(StdLink {
compiler: from,
target_compiler: compiler,
target: target,
target,
});
return;
}
@ -100,7 +100,7 @@ impl Step for Std {
builder.ensure(StdLink {
compiler: builder.compiler(compiler.stage, build.build),
target_compiler: compiler,
target: target,
target,
});
}
}
@ -202,7 +202,7 @@ impl Step for StdLink {
builder.ensure(tool::CleanTools {
compiler: target_compiler,
target: target,
target,
mode: Mode::Libstd,
});
}
@ -326,13 +326,13 @@ impl Step for Test {
if build.force_use_stage1(compiler, target) {
builder.ensure(Test {
compiler: builder.compiler(1, build.build),
target: target,
target,
});
println!("Uplifting stage1 test ({} -> {})", &build.build, target);
builder.ensure(TestLink {
compiler: builder.compiler(1, build.build),
target_compiler: compiler,
target: target,
target,
});
return;
}
@ -351,7 +351,7 @@ impl Step for Test {
builder.ensure(TestLink {
compiler: builder.compiler(compiler.stage, build.build),
target_compiler: compiler,
target: target,
target,
});
}
}
@ -398,7 +398,7 @@ impl Step for TestLink {
&libtest_stamp(build, compiler, target));
builder.ensure(tool::CleanTools {
compiler: target_compiler,
target: target,
target,
mode: Mode::Libtest,
});
}
@ -445,7 +445,7 @@ impl Step for Rustc {
if build.force_use_stage1(compiler, target) {
builder.ensure(Rustc {
compiler: builder.compiler(1, build.build),
target: target,
target,
});
println!("Uplifting stage1 rustc ({} -> {})", &build.build, target);
builder.ensure(RustcLink {
@ -581,7 +581,7 @@ impl Step for RustcLink {
&librustc_stamp(build, compiler, target));
builder.ensure(tool::CleanTools {
compiler: target_compiler,
target: target,
target,
mode: Mode::Librustc,
});
}

View File

@ -213,13 +213,13 @@ impl Step for TheBook {
let name = self.name;
// build book first edition
builder.ensure(Rustbook {
target: target,
target,
name: INTERNER.intern_string(format!("{}/first-edition", name)),
});
// build book second edition
builder.ensure(Rustbook {
target: target,
target,
name: INTERNER.intern_string(format!("{}/second-edition", name)),
});

View File

@ -274,14 +274,14 @@ Arguments:
}
"test" => {
Subcommand::Test {
paths: paths,
paths,
test_args: matches.opt_strs("test-args"),
fail_fast: !matches.opt_present("no-fail-fast"),
}
}
"bench" => {
Subcommand::Bench {
paths: paths,
paths,
test_args: matches.opt_strs("test-args"),
}
}
@ -297,12 +297,12 @@ Arguments:
}
"dist" => {
Subcommand::Dist {
paths: paths,
paths,
}
}
"install" => {
Subcommand::Install {
paths: paths,
paths,
}
}
_ => {
@ -324,7 +324,7 @@ Arguments:
Flags {
verbose: matches.opt_count("verbose"),
stage: stage,
stage,
on_fail: matches.opt_str("on-fail"),
keep_stage: matches.opt_str("keep-stage").map(|j| j.parse().unwrap()),
build: matches.opt_str("build").map(|s| INTERNER.intern_string(s)),
@ -333,9 +333,9 @@ Arguments:
target: split(matches.opt_strs("target"))
.into_iter().map(|x| INTERNER.intern_string(x)).collect::<Vec<_>>(),
config: cfg_file,
src: src,
src,
jobs: matches.opt_str("jobs").map(|j| j.parse().unwrap()),
cmd: cmd,
cmd,
incremental: matches.opt_present("incremental"),
}
}

View File

@ -314,19 +314,19 @@ impl Build {
hosts: config.hosts.clone(),
targets: config.targets.clone(),
config: config,
src: src,
out: out,
config,
src,
out,
rust_info: rust_info,
cargo_info: cargo_info,
rls_info: rls_info,
rust_info,
cargo_info,
rls_info,
cc: HashMap::new(),
cxx: HashMap::new(),
crates: HashMap::new(),
lldb_version: None,
lldb_python_dir: None,
is_sudo: is_sudo,
is_sudo,
ci_env: CiEnv::current(),
delayed_failures: Cell::new(0),
}

View File

@ -75,10 +75,10 @@ fn build_krate(build: &mut Build, krate: &str) {
doc_step: format!("doc-crate-{}", name),
test_step: format!("test-crate-{}", name),
bench_step: format!("bench-crate-{}", name),
name: name,
name,
version: package.version,
deps: Vec::new(),
path: path,
path,
});
}
}

View File

@ -278,7 +278,7 @@ impl<T> Arc<T> {
let x: Box<_> = box ArcInner {
strong: atomic::AtomicUsize::new(1),
weak: atomic::AtomicUsize::new(1),
data: data,
data,
};
Arc { ptr: Shared::from(Box::into_unique(x)) }
}

View File

@ -853,9 +853,9 @@ impl<'a, T> Hole<'a, T> {
debug_assert!(pos < data.len());
let elt = ptr::read(&data[pos]);
Hole {
data: data,
data,
elt: Some(elt),
pos: pos,
pos,
}
}
@ -1203,7 +1203,7 @@ where T: Clone + Ord {
let place = Placer::make_place(self.data.place_back());
BinaryHeapPlace {
heap: ptr,
place: place,
place,
}
}
}

View File

@ -169,7 +169,7 @@ fn make_place<T>() -> IntermediateBox<T> {
IntermediateBox {
ptr: p,
layout: layout,
layout,
marker: marker::PhantomData,
}
}

View File

@ -234,7 +234,7 @@ impl<K, Q: ?Sized> super::Recover<Q> for BTreeMap<K, ()>
match search::search_tree(self.root.as_mut(), key) {
Found(handle) => {
Some(OccupiedEntry {
handle: handle,
handle,
length: &mut self.length,
_marker: PhantomData,
}
@ -250,8 +250,8 @@ impl<K, Q: ?Sized> super::Recover<Q> for BTreeMap<K, ()>
Found(handle) => Some(mem::replace(handle.into_kv_mut().0, key)),
GoDown(handle) => {
VacantEntry {
key: key,
handle: handle,
key,
handle,
length: &mut self.length,
_marker: PhantomData,
}
@ -695,7 +695,7 @@ impl<K: Ord, V> BTreeMap<K, V> {
match search::search_tree(self.root.as_mut(), key) {
Found(handle) => {
Some(OccupiedEntry {
handle: handle,
handle,
length: &mut self.length,
_marker: PhantomData,
}
@ -866,15 +866,15 @@ impl<K: Ord, V> BTreeMap<K, V> {
match search::search_tree(self.root.as_mut(), &key) {
Found(handle) => {
Occupied(OccupiedEntry {
handle: handle,
handle,
length: &mut self.length,
_marker: PhantomData,
})
}
GoDown(handle) => {
Vacant(VacantEntry {
key: key,
handle: handle,
key,
handle,
length: &mut self.length,
_marker: PhantomData,
})

View File

@ -776,8 +776,8 @@ impl<BorrowType, K, V, NodeType> Handle<NodeRef<BorrowType, K, V, NodeType>, mar
debug_assert!(idx < node.len());
Handle {
node: node,
idx: idx,
node,
idx,
_marker: PhantomData
}
}
@ -850,8 +850,8 @@ impl<BorrowType, K, V, NodeType>
debug_assert!(idx <= node.len());
Handle {
node: node,
idx: idx,
node,
idx,
_marker: PhantomData
}
}
@ -1149,7 +1149,7 @@ impl<'a, K, V> Handle<NodeRef<marker::Mut<'a>, K, V, marker::Internal>, marker::
let mut new_root = Root {
node: BoxedNode::from_internal(new_node),
height: height
height,
};
for i in 0..(new_len+1) {
@ -1449,12 +1449,12 @@ impl<BorrowType, K, V, HandleType>
> {
match self.node.force() {
ForceResult::Leaf(node) => ForceResult::Leaf(Handle {
node: node,
node,
idx: self.idx,
_marker: PhantomData
}),
ForceResult::Internal(node) => ForceResult::Internal(Handle {
node: node,
node,
idx: self.idx,
_marker: PhantomData
})

View File

@ -140,7 +140,7 @@ impl<T> Node<T> {
Node {
next: None,
prev: None,
element: element,
element,
}
}
@ -924,7 +924,7 @@ impl<'a, T> IterMut<'a, T> {
let node = Some(Shared::from(Box::into_unique(box Node {
next: Some(head),
prev: Some(prev),
element: element,
element,
})));
prev.as_mut().next = node;

View File

@ -60,8 +60,8 @@ impl<T, A: Alloc> RawVec<T, A> {
// Unique::empty() doubles as "unallocated" and "zero-sized allocation"
RawVec {
ptr: Unique::empty(),
cap: cap,
a: a,
cap,
a,
}
}
@ -104,8 +104,8 @@ impl<T, A: Alloc> RawVec<T, A> {
RawVec {
ptr: Unique::new_unchecked(ptr as *mut _),
cap: cap,
a: a,
cap,
a,
}
}
}
@ -159,8 +159,8 @@ impl<T, A: Alloc> RawVec<T, A> {
pub unsafe fn from_raw_parts_in(ptr: *mut T, cap: usize, a: A) -> Self {
RawVec {
ptr: Unique::new_unchecked(ptr),
cap: cap,
a: a,
cap,
a,
}
}
}
@ -176,7 +176,7 @@ impl<T> RawVec<T, Heap> {
pub unsafe fn from_raw_parts(ptr: *mut T, cap: usize) -> Self {
RawVec {
ptr: Unique::new_unchecked(ptr),
cap: cap,
cap,
a: Heap,
}
}

View File

@ -311,7 +311,7 @@ impl<T> Rc<T> {
ptr: Shared::from(Box::into_unique(box RcBox {
strong: Cell::new(1),
weak: Cell::new(1),
value: value,
value,
})),
}
}

View File

@ -1886,7 +1886,7 @@ fn merge_sort<T, F>(v: &mut [T], mut is_less: F)
// Push this run onto the stack.
runs.push(Run {
start: start,
start,
len: end - start,
});
end = start;

View File

@ -1378,8 +1378,8 @@ impl String {
let chars_iter = self[start..end].chars();
Drain {
start: start,
end: end,
start,
end,
iter: chars_iter,
string: self_ptr,
}
@ -1442,11 +1442,11 @@ impl String {
let chars_iter = self[start..end].chars();
Splice {
start: start,
end: end,
start,
end,
iter: chars_iter,
string: self_ptr,
replace_with: replace_with
replace_with,
}
}

View File

@ -1728,9 +1728,9 @@ impl<T> IntoIterator for Vec<T> {
mem::forget(self);
IntoIter {
buf: Shared::new_unchecked(begin),
cap: cap,
cap,
ptr: begin,
end: end,
end,
}
}
}

View File

@ -2442,7 +2442,7 @@ impl<T> From<Vec<T>> for VecDeque<T> {
VecDeque {
tail: 0,
head: len,
buf: buf,
buf,
}
}
}

View File

@ -18,7 +18,7 @@ struct PadAdapter<'a, 'b: 'a> {
impl<'a, 'b: 'a> PadAdapter<'a, 'b> {
fn new(fmt: &'a mut fmt::Formatter<'b>) -> PadAdapter<'a, 'b> {
PadAdapter {
fmt: fmt,
fmt,
on_newline: false,
}
}
@ -94,8 +94,8 @@ pub fn debug_struct_new<'a, 'b>(fmt: &'a mut fmt::Formatter<'b>,
-> DebugStruct<'a, 'b> {
let result = fmt.write_str(name);
DebugStruct {
fmt: fmt,
result: result,
fmt,
result,
has_fields: false,
}
}
@ -185,8 +185,8 @@ pub struct DebugTuple<'a, 'b: 'a> {
pub fn debug_tuple_new<'a, 'b>(fmt: &'a mut fmt::Formatter<'b>, name: &str) -> DebugTuple<'a, 'b> {
let result = fmt.write_str(name);
DebugTuple {
fmt: fmt,
result: result,
fmt,
result,
fields: 0,
empty_name: name.is_empty(),
}
@ -317,8 +317,8 @@ pub fn debug_set_new<'a, 'b>(fmt: &'a mut fmt::Formatter<'b>) -> DebugSet<'a, 'b
let result = write!(fmt, "{{");
DebugSet {
inner: DebugInner {
fmt: fmt,
result: result,
fmt,
result,
has_fields: false,
},
}
@ -388,8 +388,8 @@ pub fn debug_list_new<'a, 'b>(fmt: &'a mut fmt::Formatter<'b>) -> DebugList<'a,
let result = write!(fmt, "[");
DebugList {
inner: DebugInner {
fmt: fmt,
result: result,
fmt,
result,
has_fields: false,
},
}
@ -460,8 +460,8 @@ pub struct DebugMap<'a, 'b: 'a> {
pub fn debug_map_new<'a, 'b>(fmt: &'a mut fmt::Formatter<'b>) -> DebugMap<'a, 'b> {
let result = write!(fmt, "{{");
DebugMap {
fmt: fmt,
result: result,
fmt,
result,
has_fields: false,
}
}

View File

@ -334,9 +334,9 @@ impl<'a> Arguments<'a> {
pub fn new_v1(pieces: &'a [&'a str],
args: &'a [ArgumentV1<'a>]) -> Arguments<'a> {
Arguments {
pieces: pieces,
pieces,
fmt: None,
args: args
args,
}
}
@ -353,9 +353,9 @@ impl<'a> Arguments<'a> {
args: &'a [ArgumentV1<'a>],
fmt: &'a [rt::v1::Argument]) -> Arguments<'a> {
Arguments {
pieces: pieces,
pieces,
fmt: Some(fmt),
args: args
args,
}
}

View File

@ -840,8 +840,8 @@ impl<A, B> ZipImpl<A, B> for Zip<A, B>
type Item = (A::Item, B::Item);
default fn new(a: A, b: B) -> Self {
Zip {
a: a,
b: b,
a,
b,
index: 0, // unused
len: 0, // unused
}
@ -903,10 +903,10 @@ impl<A, B> ZipImpl<A, B> for Zip<A, B>
fn new(a: A, b: B) -> Self {
let len = cmp::min(a.len(), b.len());
Zip {
a: a,
b: b,
a,
b,
index: 0,
len: len,
len,
}
}

View File

@ -744,7 +744,7 @@ impl<I, T, E> ResultShunt<I, E>
fn new(iter: I) -> Self {
ResultShunt {
iter: iter,
iter,
error: None,
}
}

View File

@ -86,7 +86,7 @@ impl Fp {
assert_eq!(self.f << edelta >> edelta, self.f);
Fp {
f: self.f << edelta,
e: e,
e,
}
}
}

View File

@ -442,7 +442,7 @@ pub fn to_shortest_str<'a, T, F>(mut format_shortest: F, v: T,
}
FullDecoded::Finite(ref decoded) => {
let (len, exp) = format_shortest(decoded, buf);
Formatted { sign: sign,
Formatted { sign,
parts: digits_to_dec_str(&buf[..len], exp, frac_digits, parts) }
}
}
@ -581,7 +581,7 @@ pub fn to_exact_exp_str<'a, T, F>(mut format_exact: F, v: T,
let trunc = if ndigits < maxlen { ndigits } else { maxlen };
let (len, exp) = format_exact(decoded, &mut buf[..trunc], i16::MIN);
Formatted { sign: sign,
Formatted { sign,
parts: digits_to_exp_str(&buf[..len], exp, ndigits, upper, parts) }
}
}
@ -652,7 +652,7 @@ pub fn to_exact_fixed_str<'a, T, F>(mut format_exact: F, v: T,
Formatted { sign: sign, parts: &parts[..1] }
}
} else {
Formatted { sign: sign,
Formatted { sign,
parts: digits_to_dec_str(&buf[..len], exp, frac_digits, parts) }
}
}

View File

@ -300,7 +300,7 @@ impl<T> SliceExt for [T] {
{
Split {
v: self,
pred: pred,
pred,
finished: false
}
}

View File

@ -2233,7 +2233,7 @@ impl StrExt for str {
fn splitn<'a, P: Pattern<'a>>(&'a self, count: usize, pat: P) -> SplitN<'a, P> {
SplitN(SplitNInternal {
iter: self.split(pat).0,
count: count,
count,
})
}

View File

@ -290,7 +290,7 @@ impl<'a, C: CharEq> Pattern<'a> for CharEqPattern<C> {
fn into_searcher(self, haystack: &'a str) -> CharEqSearcher<'a, C> {
CharEqSearcher {
ascii_only: self.0.only_ascii(),
haystack: haystack,
haystack,
char_eq: self.0,
char_indices: haystack.char_indices(),
}
@ -596,8 +596,8 @@ impl<'a, 'b> StrSearcher<'a, 'b> {
fn new(haystack: &'a str, needle: &'b str) -> StrSearcher<'a, 'b> {
if needle.is_empty() {
StrSearcher {
haystack: haystack,
needle: needle,
haystack,
needle,
searcher: StrSearcherImpl::Empty(EmptyNeedle {
position: 0,
end: haystack.len(),
@ -607,8 +607,8 @@ impl<'a, 'b> StrSearcher<'a, 'b> {
}
} else {
StrSearcher {
haystack: haystack,
needle: needle,
haystack,
needle,
searcher: StrSearcherImpl::TwoWay(
TwoWaySearcher::new(needle.as_bytes(), haystack.len())
),
@ -899,13 +899,13 @@ impl TwoWaySearcher {
TwoWaySearcher::reverse_maximal_suffix(needle, period, true));
TwoWaySearcher {
crit_pos: crit_pos,
crit_pos_back: crit_pos_back,
period: period,
crit_pos,
crit_pos_back,
period,
byteset: Self::byteset_create(&needle[..period]),
position: 0,
end: end,
end,
memory: 0,
memory_back: needle.len(),
}
@ -918,13 +918,13 @@ impl TwoWaySearcher {
// reverse search.
TwoWaySearcher {
crit_pos: crit_pos,
crit_pos,
crit_pos_back: crit_pos,
period: cmp::max(crit_pos, needle.len() - crit_pos) + 1,
byteset: Self::byteset_create(needle),
position: 0,
end: end,
end,
memory: usize::MAX, // Dummy value to signify that the period is long
memory_back: usize::MAX,
}

View File

@ -392,7 +392,7 @@ pub struct CycleIter<'a, T: 'a> {
pub fn cycle<T>(data: &[T]) -> CycleIter<T> {
CycleIter {
index: 0,
data: data,
data,
}
}

View File

@ -53,7 +53,7 @@ fn test_get_resource() {
fn r(i: Rc<RefCell<isize>>) -> R {
R {
i: i
i,
}
}

View File

@ -284,7 +284,7 @@ impl<'a> Parser<'a> {
Argument {
position: pos,
format: format,
format,
}
}

View File

@ -250,28 +250,28 @@ impl OptGroup {
(0, _) => {
Opt {
name: Long((long_name)),
hasarg: hasarg,
occur: occur,
hasarg,
occur,
aliases: Vec::new(),
}
}
(1, 0) => {
Opt {
name: Short(short_name.chars().next().unwrap()),
hasarg: hasarg,
occur: occur,
hasarg,
occur,
aliases: Vec::new(),
}
}
(1, _) => {
Opt {
name: Long((long_name)),
hasarg: hasarg,
occur: occur,
hasarg,
occur,
aliases: vec![Opt {
name: Short(short_name.chars().next().unwrap()),
hasarg: hasarg,
occur: occur,
hasarg,
occur,
aliases: Vec::new(),
}],
}
@ -530,8 +530,8 @@ pub fn opt(short_name: &str,
long_name: long_name.to_owned(),
hint: hint.to_owned(),
desc: desc.to_owned(),
hasarg: hasarg,
occur: occur,
hasarg,
occur,
}
}
@ -681,9 +681,9 @@ pub fn getopts(args: &[String], optgrps: &[OptGroup]) -> Result {
}
}
Ok(Matches {
opts: opts,
vals: vals,
free: free,
opts,
vals,
free,
})
}

View File

@ -781,10 +781,10 @@ mod tests {
fn edge(from: usize, to: usize, label: &'static str, style: Style) -> Edge {
Edge {
from: from,
to: to,
label: label,
style: style,
from,
to,
label,
style,
}
}
@ -848,9 +848,9 @@ mod tests {
-> LabelledGraph {
let count = node_labels.len();
LabelledGraph {
name: name,
name,
node_labels: node_labels.to_opt_strs(),
edges: edges,
edges,
node_styles: match node_styles {
Some(nodes) => nodes,
None => vec![Style::None; count],

View File

@ -74,7 +74,7 @@ pub unsafe fn panic(data: Box<Any + Send>) -> u32 {
let exception = Box::new(Exception {
_uwe: uw::_Unwind_Exception {
exception_class: rust_exception_class(),
exception_cleanup: exception_cleanup,
exception_cleanup,
private: [0; uw::unwinder_private_data_size],
},
cause: Some(data),

View File

@ -127,9 +127,9 @@ impl GammaLargeShape {
fn new_raw(shape: f64, scale: f64) -> GammaLargeShape {
let d = shape - 1. / 3.;
GammaLargeShape {
scale: scale,
scale,
c: 1. / (9. * d).sqrt(),
d: d,
d,
}
}
}

View File

@ -149,7 +149,7 @@ impl<'a, T: Clone> WeightedChoice<'a, T> {
"WeightedChoice::new called with a total weight of 0");
WeightedChoice {
items: items,
items,
// we're likely to be generating numbers in this range
// relatively often, so might as well cache it
weight_range: Range::new(0, running_total),

View File

@ -103,8 +103,8 @@ impl Normal {
pub fn new(mean: f64, std_dev: f64) -> Normal {
assert!(std_dev >= 0.0, "Normal::new called with `std_dev` < 0");
Normal {
mean: mean,
std_dev: std_dev,
mean,
std_dev,
}
}
}

View File

@ -104,7 +104,7 @@ macro_rules! integer_impl {
let zone = unsigned_max - unsigned_max % range;
Range {
low: low,
low,
range: range as $ty,
accept_zone: zone as $ty
}
@ -143,7 +143,7 @@ macro_rules! float_impl {
impl SampleRange for $ty {
fn construct_range(low: $ty, high: $ty) -> Range<$ty> {
Range {
low: low,
low,
range: high - low,
accept_zone: 0.0 // unused
}

View File

@ -418,10 +418,10 @@ impl Rand for XorShiftRng {
}
let (x, y, z, w) = tuple;
XorShiftRng {
x: x,
y: y,
z: z,
w: w,
x,
y,
z,
w,
}
}
}

View File

@ -38,10 +38,10 @@ impl<R: Rng, Rsdr: Reseeder<R>> ReseedingRng<R, Rsdr> {
/// * `reseeder`: the reseeding object to use.
pub fn new(rng: R, generation_threshold: usize, reseeder: Rsdr) -> ReseedingRng<R, Rsdr> {
ReseedingRng {
rng: rng,
generation_threshold: generation_threshold,
rng,
generation_threshold,
bytes_generated: 0,
reseeder: reseeder,
reseeder,
}
}

View File

@ -630,8 +630,8 @@ pub fn check_crate<'a, 'tcx>(tcx: TyCtxt<'a, 'tcx, 'tcx>) {
let krate = tcx.hir.krate();
let live_symbols = find_live(tcx, access_levels, krate);
let mut visitor = DeadVisitor {
tcx: tcx,
live_symbols: live_symbols,
tcx,
live_symbols,
};
intravisit::walk_crate(&mut visitor, krate);
}

View File

@ -1695,8 +1695,8 @@ impl<'a, 'gcx, 'tcx> TyCtxt<'a, 'gcx, 'tcx> {
substs: &'tcx Substs<'tcx>)
-> Ty<'tcx> {
self.mk_ty(TyProjection(ProjectionTy {
item_def_id: item_def_id,
substs: substs,
item_def_id,
substs,
}))
}

View File

@ -253,7 +253,7 @@ impl<'tcx> Relate<'tcx> for ty::ExistentialProjection<'tcx> {
let substs = relation.relate(&a.substs, &b.substs)?;
Ok(ty::ExistentialProjection {
item_def_id: a.item_def_id,
substs: substs,
substs,
ty,
})
}

View File

@ -137,7 +137,7 @@ impl<'a, 'tcx> Lift<'tcx> for ty::ProjectionTy<'a> {
tcx.lift(&self.substs).map(|substs| {
ty::ProjectionTy {
item_def_id: self.item_def_id,
substs: substs,
substs,
}
})
}

View File

@ -569,7 +569,7 @@ impl<'a, 'tcx> ProjectionTy<'tcx> {
pub fn trait_ref(&self, tcx: TyCtxt) -> ty::TraitRef<'tcx> {
let def_id = tcx.associated_item(self.item_def_id).container.id();
ty::TraitRef {
def_id: def_id,
def_id,
substs: self.substs,
}
}
@ -874,7 +874,7 @@ impl<'a, 'tcx, 'gcx> ExistentialProjection<'tcx> {
pub fn trait_ref(&self, tcx: TyCtxt) -> ty::ExistentialTraitRef<'tcx> {
let def_id = tcx.associated_item(self.item_def_id).container.id();
ty::ExistentialTraitRef{
def_id: def_id,
def_id,
substs: self.substs,
}
}

View File

@ -36,9 +36,9 @@ pub fn modify(sess: &ParseSess,
krate: Crate,
handler: &rustc_errors::Handler) -> ast::Crate {
ExpandAllocatorDirectives {
handler: handler,
sess: sess,
resolver: resolver,
handler,
sess,
resolver,
found: false,
}.fold_crate(krate)
}
@ -88,7 +88,7 @@ impl<'a> Folder for ExpandAllocatorDirectives<'a> {
};
let ecfg = ExpansionConfig::default(name.to_string());
let mut f = AllocFnFactory {
span: span,
span,
kind: AllocatorKind::Global,
global: item.ident,
alloc: Ident::from_str("alloc"),

View File

@ -97,7 +97,7 @@ pub fn opts(arch: Arch) -> Result<TargetOptions, String> {
cpu: target_cpu(arch),
dynamic_linking: false,
executables: true,
pre_link_args: pre_link_args,
pre_link_args,
has_elf_tls: false,
.. super::apple_base::opts()
})

View File

@ -25,7 +25,7 @@ pub fn opts() -> TargetOptions {
exe_allocation_crate: Some("alloc_system".to_string()),
panic_strategy: PanicStrategy::Abort,
linker: "ld".to_string(),
pre_link_args: pre_link_args,
pre_link_args,
target_family: Some("unix".to_string()),
.. Default::default()
}

View File

@ -25,8 +25,8 @@ pub fn target() -> TargetResult {
linker: "pnacl-clang".to_string(),
ar: "pnacl-ar".to_string(),
pre_link_args: pre_link_args,
post_link_args: post_link_args,
pre_link_args,
post_link_args,
dynamic_linking: false,
executables: true,
exe_suffix: ".pexe".to_string(),

View File

@ -38,7 +38,7 @@ pub fn target() -> Result<Target, String> {
obj_is_bitcode: true,
is_like_emscripten: true,
max_atomic_width: Some(32),
post_link_args: post_link_args,
post_link_args,
target_family: Some("unix".to_string()),
.. Default::default()
};

View File

@ -34,7 +34,7 @@ pub fn target() -> Result<Target, String> {
obj_is_bitcode: true,
is_like_emscripten: true,
max_atomic_width: Some(32),
post_link_args: post_link_args,
post_link_args,
target_family: Some("unix".to_string()),
.. Default::default()
};

View File

@ -78,7 +78,7 @@ pub fn opts() -> TargetOptions {
target_family: Some("windows".to_string()),
is_like_windows: true,
allows_weak_linkage: false,
pre_link_args: pre_link_args,
pre_link_args,
pre_link_objects_exe: vec![
"crt2.o".to_string(), // mingw C runtime initialization for executables
"rsbegin.o".to_string(), // Rust compiler runtime initialization, see rsbegin.rs
@ -87,7 +87,7 @@ pub fn opts() -> TargetOptions {
"dllcrt2.o".to_string(), // mingw C runtime initialization for dlls
"rsbegin.o".to_string(),
],
late_link_args: late_link_args,
late_link_args,
post_link_objects: vec![
"rsend.o".to_string()
],

View File

@ -194,10 +194,10 @@ pub fn check_loans<'a, 'b, 'c, 'tcx>(bccx: &BorrowckCtxt<'a, 'tcx>,
let def_id = bccx.tcx.hir.body_owner_def_id(body.id());
let param_env = bccx.tcx.param_env(def_id);
let mut clcx = CheckLoanCtxt {
bccx: bccx,
dfcx_loans: dfcx_loans,
move_data: move_data,
all_loans: all_loans,
bccx,
dfcx_loans,
move_data,
all_loans,
param_env,
};
euv::ExprUseVisitor::new(&mut clcx, bccx.tcx, param_env, &bccx.region_maps, bccx.tables)

View File

@ -106,8 +106,8 @@ pub fn gather_move_from_expr<'a, 'tcx>(bccx: &BorrowckCtxt<'a, 'tcx>,
};
let move_info = GatherMoveInfo {
id: move_expr_id,
kind: kind,
cmt: cmt,
kind,
cmt,
span_path_opt: None,
};
gather_move(bccx, move_data, move_error_collector, move_info);
@ -163,7 +163,7 @@ pub fn gather_move_from_pat<'a, 'tcx>(bccx: &BorrowckCtxt<'a, 'tcx>,
let move_info = GatherMoveInfo {
id: move_pat.id,
kind: MovePat,
cmt: cmt,
cmt,
span_path_opt: pat_span_path_opt,
};

View File

@ -37,10 +37,10 @@ pub fn guarantee_lifetime<'a, 'tcx>(bccx: &BorrowckCtxt<'a, 'tcx>,
debug!("guarantee_lifetime(cmt={:?}, loan_region={:?})",
cmt, loan_region);
let ctxt = GuaranteeLifetimeContext {bccx: bccx,
item_scope: item_scope,
span: span,
cause: cause,
loan_region: loan_region,
item_scope,
span,
cause,
loan_region,
cmt_original: cmt.clone()};
ctxt.check(&cmt, None)
}

View File

@ -41,7 +41,7 @@ pub fn gather_loans_in_fn<'a, 'tcx>(bccx: &BorrowckCtxt<'a, 'tcx>,
let def_id = bccx.tcx.hir.body_owner_def_id(body);
let param_env = bccx.tcx.param_env(def_id);
let mut glcx = GatherLoanCtxt {
bccx: bccx,
bccx,
all_loans: Vec::new(),
item_ub: region::CodeExtent::Misc(body.node_id),
move_data: MoveData::new(),
@ -230,8 +230,8 @@ fn check_mutability<'a, 'tcx>(bccx: &BorrowckCtxt<'a, 'tcx>,
// Only mutable data can be lent as mutable.
if !cmt.mutbl.is_mutable() {
Err(bccx.report(BckError { span: borrow_span,
cause: cause,
cmt: cmt,
cause,
cmt,
code: err_mutbl }))
} else {
Ok(())
@ -389,13 +389,13 @@ impl<'a, 'tcx> GatherLoanCtxt<'a, 'tcx> {
Loan {
index: self.all_loans.len(),
loan_path: loan_path,
loan_path,
kind: req_kind,
gen_scope: gen_scope,
kill_scope: kill_scope,
gen_scope,
kill_scope,
span: borrow_span,
restricted_paths: restricted_paths,
cause: cause,
restricted_paths,
cause,
}
}
};
@ -423,13 +423,13 @@ impl<'a, 'tcx> GatherLoanCtxt<'a, 'tcx> {
// let all_loans = &mut *self.all_loans; // FIXME(#5074)
// Loan {
// index: all_loans.len(),
// loan_path: loan_path,
// cmt: cmt,
// loan_path,
// cmt,
// mutbl: ConstMutability,
// gen_scope: borrow_id,
// kill_scope: kill_scope,
// kill_scope,
// span: borrow_span,
// restrictions: restrictions
// restrictions,
// }
// }
}

View File

@ -49,8 +49,8 @@ impl<'tcx> MoveError<'tcx> {
move_to: Option<MovePlace<'tcx>>)
-> MoveError<'tcx> {
MoveError {
move_from: move_from,
move_to: move_to,
move_from,
move_to,
}
}
}

View File

@ -34,10 +34,10 @@ pub fn compute_restrictions<'a, 'tcx>(bccx: &BorrowckCtxt<'a, 'tcx>,
loan_region: ty::Region<'tcx>)
-> RestrictionResult<'tcx> {
let ctxt = RestrictionsContext {
bccx: bccx,
span: span,
cause: cause,
loan_region: loan_region,
bccx,
span,
cause,
loan_region,
};
ctxt.restrict(cmt)

View File

@ -180,7 +180,7 @@ fn build_borrowck_dataflow_data<'a, 'c, 'tcx, F>(this: &mut BorrowckCtxt<'a, 'tc
id_range,
body);
Some(AnalysisData { all_loans: all_loans,
Some(AnalysisData { all_loans,
loans: loan_dfcx,
move_data:flowed_moves })
}

View File

@ -310,7 +310,7 @@ impl<'a, 'tcx> MoveData<'tcx> {
parent: parent_index,
first_move: InvalidMoveIndex,
first_child: InvalidMovePathIndex,
next_sibling: next_sibling,
next_sibling,
});
index
@ -408,9 +408,9 @@ impl<'a, 'tcx> MoveData<'tcx> {
self.moves.borrow_mut().push(Move {
path: path_index,
id: id,
kind: kind,
next_move: next_move
id,
kind,
next_move,
});
}
@ -468,8 +468,8 @@ impl<'a, 'tcx> MoveData<'tcx> {
let assignment = Assignment {
path: path_index,
id: assign_id,
span: span,
assignee_id: assignee_id,
span,
assignee_id,
};
if self.is_var_path(path_index) {
@ -504,7 +504,7 @@ impl<'a, 'tcx> MoveData<'tcx> {
path: path_index,
base_path: base_path_index,
id: pattern_id,
mode: mode,
mode,
};
self.variant_matches.borrow_mut().push(variant_match);
@ -680,9 +680,9 @@ impl<'a, 'tcx> FlowedMoveData<'a, 'tcx> {
dfcx_assign.propagate(cfg, body);
FlowedMoveData {
move_data: move_data,
dfcx_moves: dfcx_moves,
dfcx_assign: dfcx_assign,
move_data,
dfcx_moves,
dfcx_assign,
}
}

View File

@ -166,8 +166,8 @@ impl<'a, 'tcx> MatchCheckCtxt<'a, 'tcx> {
let pattern_arena = TypedArena::new();
f(MatchCheckCtxt {
tcx: tcx,
module: module,
tcx,
module,
pattern_arena: &pattern_arena,
byte_array_map: FxHashMap(),
})
@ -296,7 +296,7 @@ impl<'tcx> Witness<'tcx> {
let sub_pattern_tys = constructor_sub_pattern_tys(cx, ctor, ty);
self.0.extend(sub_pattern_tys.into_iter().map(|ty| {
Pattern {
ty: ty,
ty,
span: DUMMY_SP,
kind: box PatternKind::Wild,
}
@ -344,7 +344,7 @@ impl<'tcx> Witness<'tcx> {
if adt.variants.len() > 1 {
PatternKind::Variant {
adt_def: adt,
substs: substs,
substs,
variant_index: ctor.variant_index_for_adt(adt),
subpatterns: pats
}
@ -378,7 +378,7 @@ impl<'tcx> Witness<'tcx> {
};
self.0.push(Pattern {
ty: ty,
ty,
span: DUMMY_SP,
kind: Box::new(pat),
});
@ -673,7 +673,7 @@ fn is_useful_specialized<'p, 'a:'p, 'tcx: 'a>(
let sub_pat_tys = constructor_sub_pattern_tys(cx, &ctor, lty);
let wild_patterns_owned: Vec<_> = sub_pat_tys.iter().map(|ty| {
Pattern {
ty: ty,
ty,
span: DUMMY_SP,
kind: box PatternKind::Wild,
}

View File

@ -524,7 +524,7 @@ fn check_legality_of_move_bindings(cx: &MatchVisitor,
/// FIXME: this should be done by borrowck.
fn check_for_mutation_in_guard(cx: &MatchVisitor, guard: &hir::Expr) {
let mut checker = MutationChecker {
cx: cx,
cx,
};
ExprUseVisitor::new(&mut checker, cx.tcx, cx.param_env, cx.region_maps, cx.tables)
.walk_expr(guard);

View File

@ -379,7 +379,7 @@ fn eval_const_expr_partial<'a, 'tcx>(cx: &ConstContext<'a, 'tcx>,
tcx,
param_env: cx.param_env,
tables: tcx.typeck_tables_of(def_id),
substs: substs,
substs,
fn_args: Some(call_args)
};
callee_cx.eval(&body.value)?

View File

@ -407,8 +407,8 @@ impl<'a, 'tcx> PatternContext<'a, 'tcx> {
}
PatternKind::Binding {
mutability: mutability,
mode: mode,
mutability,
mode,
name: ident.node,
var: id,
ty: var_ty,
@ -470,7 +470,7 @@ impl<'a, 'tcx> PatternContext<'a, 'tcx> {
Pattern {
span: pat.span,
ty: ty,
ty,
kind: Box::new(kind),
}
}
@ -569,10 +569,10 @@ impl<'a, 'tcx> PatternContext<'a, 'tcx> {
_ => bug!("inappropriate type for def: {:?}", ty.sty),
};
PatternKind::Variant {
adt_def: adt_def,
substs: substs,
adt_def,
substs,
variant_index: adt_def.variant_index_with_id(variant_id),
subpatterns: subpatterns,
subpatterns,
}
} else {
PatternKind::Leaf { subpatterns: subpatterns }
@ -626,8 +626,8 @@ impl<'a, 'tcx> PatternContext<'a, 'tcx> {
};
Pattern {
span: span,
ty: ty,
span,
ty,
kind: Box::new(kind),
}
}
@ -762,7 +762,7 @@ impl<'a, 'tcx> PatternContext<'a, 'tcx> {
};
Pattern {
span: span,
span,
ty: pat_ty,
kind: Box::new(kind),
}

View File

@ -281,8 +281,8 @@ impl<A: Array> IntoIterator for ArrayVec<A> {
let indices = 0..self.count;
mem::forget(self);
Iter {
indices: indices,
store: store,
indices,
store,
}
}
}

View File

@ -151,7 +151,7 @@ impl BitMatrix {
// element. Round up to an even number of u64s.
let u64s_per_row = u64s(columns);
BitMatrix {
columns: columns,
columns,
vector: vec![0; rows * u64s_per_row],
}
}

View File

@ -71,8 +71,8 @@ pub fn dominators_given_rpo<G: ControlFlowGraph>(graph: &G,
}
Dominators {
post_order_rank: post_order_rank,
immediate_dominators: immediate_dominators,
post_order_rank,
immediate_dominators,
}
}
@ -181,7 +181,7 @@ impl<Node: Idx> Dominators<Node> {
}
DominatorTree {
root: root.unwrap(),
children: children,
children,
}
}
}

View File

@ -26,7 +26,7 @@ impl TestGraph {
pub fn new(start_node: usize, edges: &[(usize, usize)]) -> Self {
let mut graph = TestGraph {
num_nodes: start_node + 1,
start_node: start_node,
start_node,
successors: HashMap::new(),
predecessors: HashMap::new(),
};

View File

@ -23,8 +23,8 @@ impl<G: ControlFlowGraph> TransposedGraph<G> {
pub fn with_start(base_graph: G, start_node: G::Node) -> Self {
TransposedGraph {
base_graph: base_graph,
start_node: start_node,
base_graph,
start_node,
}
}
}

View File

@ -153,7 +153,7 @@ impl<N: Debug, E: Debug> Graph<N, E> {
let idx = self.next_node_index();
self.nodes.push(Node {
first_edge: [INVALID_EDGE_INDEX, INVALID_EDGE_INDEX],
data: data,
data,
});
idx
}
@ -189,9 +189,9 @@ impl<N: Debug, E: Debug> Graph<N, E> {
// as the next pointers
self.edges.push(Edge {
next_edge: [source_first, target_first],
source: source,
target: target,
data: data,
source,
target,
data,
});
// adjust the firsts for each node target be the next object.
@ -269,7 +269,7 @@ impl<N: Debug, E: Debug> Graph<N, E> {
let first_edge = self.node(source).first_edge[direction.repr];
AdjacentEdges {
graph: self,
direction: direction,
direction,
next: first_edge,
}
}
@ -482,10 +482,10 @@ impl<'g, N: Debug, E: Debug> DepthFirstTraversal<'g, N, E> {
pub fn new(graph: &'g Graph<N, E>, direction: Direction) -> Self {
let visited = BitVector::new(graph.len_nodes());
DepthFirstTraversal {
graph: graph,
graph,
stack: vec![],
visited: visited,
direction: direction,
visited,
direction,
}
}
@ -496,10 +496,10 @@ impl<'g, N: Debug, E: Debug> DepthFirstTraversal<'g, N, E> {
let mut visited = BitVector::new(graph.len_nodes());
visited.insert(start_node.node_id());
DepthFirstTraversal {
graph: graph,
graph,
stack: vec![start_node],
visited: visited,
direction: direction,
visited,
direction,
}
}

View File

@ -269,7 +269,7 @@ impl<O: ForestObligation> ObligationForest<O> {
let backtrace = self.error_at(index);
errors.push(Error {
error: error.clone(),
backtrace: backtrace,
backtrace,
});
}
}
@ -346,7 +346,7 @@ impl<O: ForestObligation> ObligationForest<O> {
let backtrace = self.error_at(index);
errors.push(Error {
error: err,
backtrace: backtrace,
backtrace,
});
}
}
@ -357,8 +357,8 @@ impl<O: ForestObligation> ObligationForest<O> {
// changed.
return Outcome {
completed: vec![],
errors: errors,
stalled: stalled,
errors,
stalled,
};
}
@ -372,8 +372,8 @@ impl<O: ForestObligation> ObligationForest<O> {
Outcome {
completed: completed_obligations,
errors: errors,
stalled: stalled,
errors,
stalled,
}
}
@ -638,8 +638,8 @@ impl<O: ForestObligation> ObligationForest<O> {
impl<O> Node<O> {
fn new(parent: Option<NodeIndex>, obligation: O) -> Node<O> {
Node {
obligation: obligation,
parent: parent,
obligation,
parent,
state: Cell::new(NodeState::Pending),
dependents: vec![],
}

View File

@ -87,8 +87,8 @@ impl<K: UnifyKey> VarValue<K> {
fn new(parent: K, value: K::Value, rank: u32) -> VarValue<K> {
VarValue {
parent: parent, // this is a root
value: value,
rank: rank,
value,
rank,
}
}
@ -98,8 +98,8 @@ impl<K: UnifyKey> VarValue<K> {
fn root(self, rank: u32, value: K::Value) -> VarValue<K> {
VarValue {
rank: rank,
value: value,
rank,
value,
..self
}
}

View File

@ -401,8 +401,8 @@ impl<'a, 'tcx> CompileState<'a, 'tcx> {
out_dir: &'a Option<PathBuf>)
-> Self {
CompileState {
input: input,
session: session,
input,
session,
out_dir: out_dir.as_ref().map(|s| &**s),
out_file: None,
arena: None,
@ -868,7 +868,7 @@ pub fn phase_2_configure_and_expand<F>(sess: &Session,
trait_map: resolver.trait_map,
maybe_unused_trait_imports: resolver.maybe_unused_trait_imports,
},
hir_forest: hir_forest,
hir_forest,
})
}

View File

@ -802,7 +802,7 @@ impl RustcDefaultCalls {
let mut cfgs = Vec::new();
for &(name, ref value) in sess.parse_sess.config.iter() {
let gated_cfg = GatedCfg::gate(&ast::MetaItem {
name: name,
name,
node: ast::MetaItemKind::Word,
span: DUMMY_SP,
});

View File

@ -174,7 +174,7 @@ impl PpSourceMode {
match *self {
PpmNormal | PpmEveryBodyLoops | PpmExpanded => {
let annotation = NoAnn {
sess: sess,
sess,
hir_map: hir_map.map(|m| m.clone()),
};
f(&annotation)
@ -182,14 +182,14 @@ impl PpSourceMode {
PpmIdentified | PpmExpandedIdentified => {
let annotation = IdentifiedAnnotation {
sess: sess,
sess,
hir_map: hir_map.map(|m| m.clone()),
};
f(&annotation)
}
PpmExpandedHygiene => {
let annotation = HygieneAnnotation {
sess: sess,
sess,
};
f(&annotation)
}
@ -211,7 +211,7 @@ impl PpSourceMode {
match *self {
PpmNormal => {
let annotation = NoAnn {
sess: sess,
sess,
hir_map: Some(hir_map.clone()),
};
f(&annotation, hir_map.forest.krate())
@ -219,7 +219,7 @@ impl PpSourceMode {
PpmIdentified => {
let annotation = IdentifiedAnnotation {
sess: sess,
sess,
hir_map: Some(hir_map.clone()),
};
f(&annotation, hir_map.forest.krate())
@ -235,7 +235,7 @@ impl PpSourceMode {
|tcx, _, _, _| {
let empty_tables = ty::TypeckTables::empty(None);
let annotation = TypedAnnotation {
tcx: tcx,
tcx,
tables: Cell::new(&empty_tables)
};
let _ignore = tcx.dep_graph.in_ignore();
@ -680,7 +680,7 @@ impl fold::Folder for ReplaceBodyWithLoop {
})
.into_iter()
.collect(),
rules: rules,
rules,
id: ast::DUMMY_NODE_ID,
span: syntax_pos::DUMMY_SP,
})
@ -739,7 +739,7 @@ fn print_flowgraph<'a, 'tcx, W: Write>(variants: Vec<borrowck_dot::Variant>,
hir_map: &tcx.hir,
cfg: &cfg,
name: format!("node_{}", code.id()),
labelled_edges: labelled_edges,
labelled_edges,
};
match code {
@ -758,7 +758,7 @@ fn print_flowgraph<'a, 'tcx, W: Write>(variants: Vec<borrowck_dot::Variant>,
let lcfg = borrowck_dot::DataflowLabeller {
inner: lcfg,
variants: variants,
variants,
borrowck_ctxt: &bccx,
analysis_data: &analysis_data,
};

View File

@ -83,9 +83,9 @@ impl Diagnostic {
pub fn new_with_code(level: Level, code: Option<String>, message: &str) -> Self {
Diagnostic {
level: level,
level,
message: vec![(message.to_owned(), Style::NoStyle)],
code: code,
code,
span: MultiSpan::new(),
children: vec![],
suggestions: vec![],
@ -306,10 +306,10 @@ impl Diagnostic {
span: MultiSpan,
render_span: Option<RenderSpan>) {
let sub = SubDiagnostic {
level: level,
level,
message: vec![(message.to_owned(), Style::NoStyle)],
span: span,
render_span: render_span,
span,
render_span,
};
self.children.push(sub);
}
@ -322,10 +322,10 @@ impl Diagnostic {
span: MultiSpan,
render_span: Option<RenderSpan>) {
let sub = SubDiagnostic {
level: level,
message: message,
span: span,
render_span: render_span,
level,
message,
span,
render_span,
};
self.children.push(sub);
}

View File

@ -179,7 +179,7 @@ impl<'a> DiagnosticBuilder<'a> {
message: &str)
-> DiagnosticBuilder<'a> {
DiagnosticBuilder {
handler: handler,
handler,
diagnostic: Diagnostic::new_with_code(level, code, message)
}
}

View File

@ -120,7 +120,7 @@ impl EmitterWriter {
if color_config.use_color() {
let dst = Destination::from_stderr();
EmitterWriter {
dst: dst,
dst,
cm: code_map,
}
} else {
@ -156,7 +156,7 @@ impl EmitterWriter {
}
// We don't have a line yet, create one
slot.lines.push(Line {
line_index: line_index,
line_index,
annotations: vec![ann],
});
slot.lines.sort();
@ -165,9 +165,9 @@ impl EmitterWriter {
}
// This is the first time we're seeing the file
file_vec.push(FileWithAnnotatedLines {
file: file,
file,
lines: vec![Line {
line_index: line_index,
line_index,
annotations: vec![ann],
}],
multiline_depth: 0,

View File

@ -155,8 +155,8 @@ impl CodeSuggestion {
let lo = primary_spans.iter().map(|sp| sp.0.lo).min().unwrap();
let hi = primary_spans.iter().map(|sp| sp.0.hi).min().unwrap();
let bounding_span = Span {
lo: lo,
hi: hi,
lo,
hi,
ctxt: NO_EXPANSION,
};
let lines = cm.span_to_lines(bounding_span).unwrap();
@ -292,8 +292,8 @@ impl Handler {
Handler {
err_count: Cell::new(0),
emitter: RefCell::new(e),
can_emit_warnings: can_emit_warnings,
treat_err_as_bug: treat_err_as_bug,
can_emit_warnings,
treat_err_as_bug,
continue_after_error: Cell::new(true),
delayed_span_bug: RefCell::new(None),
}

View File

@ -76,7 +76,7 @@ pub fn assert_dep_graph<'a, 'tcx>(tcx: TyCtxt<'a, 'tcx, 'tcx>) {
// Find annotations supplied by user (if any).
let (if_this_changed, then_this_would_need) = {
let mut visitor = IfThisChanged { tcx: tcx,
let mut visitor = IfThisChanged { tcx,
if_this_changed: vec![],
then_this_would_need: vec![] };
visitor.process_attrs(ast::CRATE_NODE_ID, &tcx.hir.krate().attrs);

View File

@ -83,15 +83,15 @@ pub fn check_dirty_clean_annotations<'a, 'tcx>(tcx: TyCtxt<'a, 'tcx, 'tcx>,
debug!("query-nodes: {:?}", query.nodes());
let krate = tcx.hir.krate();
let mut dirty_clean_visitor = DirtyCleanVisitor {
tcx: tcx,
tcx,
query: &query,
dirty_inputs: dirty_inputs,
dirty_inputs,
checked_attrs: FxHashSet(),
};
krate.visit_all_item_likes(&mut dirty_clean_visitor);
let mut all_attrs = FindAllAttrs {
tcx: tcx,
tcx,
attr_names: vec![ATTR_DIRTY, ATTR_CLEAN],
found_attrs: vec![],
};
@ -243,15 +243,15 @@ pub fn check_dirty_clean_metadata<'a, 'tcx>(
tcx.dep_graph.with_ignore(||{
let krate = tcx.hir.krate();
let mut dirty_clean_visitor = DirtyCleanMetadataVisitor {
tcx: tcx,
prev_metadata_hashes: prev_metadata_hashes,
current_metadata_hashes: current_metadata_hashes,
tcx,
prev_metadata_hashes,
current_metadata_hashes,
checked_attrs: FxHashSet(),
};
intravisit::walk_crate(&mut dirty_clean_visitor, krate);
let mut all_attrs = FindAllAttrs {
tcx: tcx,
tcx,
attr_names: vec![ATTR_DIRTY_METADATA, ATTR_CLEAN_METADATA],
found_attrs: vec![],
};

View File

@ -38,8 +38,8 @@ impl<'a, 'tcx> HashContext<'a, 'tcx> {
incremental_hashes_map: &'a IncrementalHashesMap)
-> Self {
HashContext {
tcx: tcx,
incremental_hashes_map: incremental_hashes_map,
tcx,
incremental_hashes_map,
metadata_hashes: FxHashMap(),
crate_hashes: FxHashMap(),
}

View File

@ -45,7 +45,7 @@ impl<'a, 'g, N, I, O> Classify<'a, 'g, N, I, O>
{
pub(super) fn new(r: &'a mut GraphReduce<'g, N, I, O>) -> Self {
Classify {
r: r,
r,
colors: vec![Color::White; r.in_graph.len_nodes()],
stack: vec![],
dag: Dag {

View File

@ -103,8 +103,8 @@ impl<'q> Predecessors<'q> {
Predecessors {
reduced_graph: graph,
bootstrap_outputs: bootstrap_outputs,
hashes: hashes,
bootstrap_outputs,
hashes,
}
}
}

View File

@ -57,7 +57,7 @@ pub fn save_trans_partition(sess: &Session,
let work_product = WorkProduct {
cgu_name: cgu_name.to_string(),
input_hash: partition_hash,
saved_files: saved_files,
saved_files,
};
sess.dep_graph.insert_work_product(&work_product_id, work_product);

View File

@ -94,7 +94,7 @@ impl<'a> Iterator for Iter<'a> {
::last_error().map(Err)
} else {
Some(Ok(Child {
ptr: ptr,
ptr,
_data: marker::PhantomData,
}))
}

View File

@ -82,12 +82,12 @@ impl OptimizationDiagnostic {
}
OptimizationDiagnostic {
kind: kind,
kind,
pass_name: pass_name.expect("got a non-UTF8 pass name from LLVM"),
function: function,
line: line,
column: column,
filename: filename,
function,
line,
column,
filename,
message: message.expect("got a non-UTF8 OptimizationDiagnostic message from LLVM")
}
}

View File

@ -53,7 +53,7 @@ impl<'a, 'b, 'tcx> IsolatedEncoder<'a, 'b, 'tcx> {
body: lazy_body,
tables: lazy_tables,
nested_bodies: lazy_nested_bodies,
rvalue_promotable_to_static: rvalue_promotable_to_static
rvalue_promotable_to_static,
})
}
}

View File

@ -162,8 +162,8 @@ enum LoadResult {
impl<'a> CrateLoader<'a> {
pub fn new(sess: &'a Session, cstore: &'a CStore, local_crate_name: &str) -> Self {
CrateLoader {
sess: sess,
cstore: cstore,
sess,
cstore,
next_crate_num: cstore.next_crate_num(),
local_crate_name: Symbol::intern(local_crate_name),
}
@ -184,7 +184,7 @@ impl<'a> CrateLoader<'a> {
};
Some(ExternCrateInfo {
ident: i.ident.name,
name: name,
name,
id: i.id,
dep_kind: if attr::contains_name(&i.attrs, "no_link") {
DepKind::UnexportedMacrosOnly
@ -325,25 +325,25 @@ impl<'a> CrateLoader<'a> {
});
let mut cmeta = cstore::CrateMetadata {
name: name,
name,
extern_crate: Cell::new(None),
def_path_table: Rc::new(def_path_table),
exported_symbols: exported_symbols,
trait_impls: trait_impls,
exported_symbols,
trait_impls,
proc_macros: crate_root.macro_derive_registrar.map(|_| {
self.load_derive_macros(&crate_root, dylib.clone().map(|p| p.0), span)
}),
root: crate_root,
blob: metadata,
cnum_map: RefCell::new(cnum_map),
cnum: cnum,
cnum,
codemap_import_info: RefCell::new(vec![]),
attribute_cache: RefCell::new([Vec::new(), Vec::new()]),
dep_kind: Cell::new(dep_kind),
source: cstore::CrateSource {
dylib: dylib,
rlib: rlib,
rmeta: rmeta,
dylib,
rlib,
rmeta,
},
// Initialize this with an empty set. The field is populated below
// after we were able to deserialize its contents.
@ -388,14 +388,14 @@ impl<'a> CrateLoader<'a> {
info!("falling back to a load");
let mut locate_ctxt = locator::Context {
sess: self.sess,
span: span,
ident: ident,
span,
ident,
crate_name: name,
hash: hash.map(|a| &*a),
filesearch: self.sess.target_filesearch(path_kind),
target: &self.sess.target.target,
triple: &self.sess.opts.target_triple,
root: root,
root,
rejected_via_hash: vec![],
rejected_via_triple: vec![],
rejected_via_kind: vec![],
@ -547,7 +547,7 @@ impl<'a> CrateLoader<'a> {
let mut target_only = false;
let mut locate_ctxt = locator::Context {
sess: self.sess,
span: span,
span,
ident: info.ident,
crate_name: info.name,
hash: None,
@ -596,9 +596,9 @@ impl<'a> CrateLoader<'a> {
};
ExtensionCrate {
metadata: metadata,
metadata,
dylib: dylib.map(|p| p.0),
target_only: target_only,
target_only,
}
}
@ -1221,9 +1221,9 @@ impl<'a> CrateLoader<'a> {
.collect();
let lib = NativeLibrary {
name: n,
kind: kind,
cfg: cfg,
foreign_items: foreign_items,
kind,
cfg,
foreign_items,
};
register_native_lib(self.sess, self.cstore, Some(m.span), lib);
}

View File

@ -114,7 +114,7 @@ impl CStore {
statically_included_foreign_items: RefCell::new(FxHashSet()),
dllimport_foreign_items: RefCell::new(FxHashSet()),
visible_parent_map: RefCell::new(FxHashMap()),
metadata_loader: metadata_loader,
metadata_loader,
}
}

View File

@ -282,7 +282,7 @@ impl CrateStore for cstore::CStore {
{
self.get_crate_data(cnum).root.plugin_registrar_fn.map(|index| DefId {
krate: cnum,
index: index
index,
})
}
@ -290,7 +290,7 @@ impl CrateStore for cstore::CStore {
{
self.get_crate_data(cnum).root.macro_derive_registrar.map(|index| DefId {
krate: cnum,
index: index
index,
})
}

View File

@ -69,7 +69,7 @@ pub trait Metadata<'a, 'tcx>: Copy {
opaque: opaque::Decoder::new(self.raw_bytes(), pos),
cdata: self.cdata(),
sess: self.sess().or(tcx.map(|tcx| tcx.sess)),
tcx: tcx,
tcx,
last_filemap_index: 0,
lazy_state: LazyState::NoNode,
}
@ -468,7 +468,7 @@ impl<'a, 'tcx> CrateMetadata {
fn local_def_id(&self, index: DefIndex) -> DefId {
DefId {
krate: self.cnum,
index: index,
index,
}
}
@ -703,7 +703,7 @@ impl<'a, 'tcx> CrateMetadata {
for child_index in child.children.decode((self, sess)) {
if let Some(def) = self.get_def(child_index) {
callback(def::Export {
def: def,
def,
ident: Ident::with_empty_ctxt(self.item_name(child_index)),
span: self.entry(child_index).span.decode((self, sess)),
});
@ -835,8 +835,8 @@ impl<'a, 'tcx> CrateMetadata {
};
ty::AssociatedItem {
name: name,
kind: kind,
name,
kind,
vis: item.visibility.decode(self),
defaultness: container.defaultness(),
def_id: self.local_def_id(id),

View File

@ -422,16 +422,16 @@ impl<'a, 'tcx> EncodeContext<'a, 'tcx> {
None
},
crate_deps: crate_deps,
dylib_dependency_formats: dylib_dependency_formats,
lang_items: lang_items,
lang_items_missing: lang_items_missing,
native_libraries: native_libraries,
codemap: codemap,
def_path_table: def_path_table,
impls: impls,
exported_symbols: exported_symbols,
index: index,
crate_deps,
dylib_dependency_formats,
lang_items,
lang_items_missing,
native_libraries,
codemap,
def_path_table,
impls,
exported_symbols,
index,
});
let total_bytes = self.position();
@ -719,15 +719,15 @@ impl<'a, 'b: 'a, 'tcx: 'b> IsolatedEncoder<'a, 'b, 'tcx> {
};
FnData {
constness: hir::Constness::NotConst,
arg_names: arg_names,
arg_names,
sig: self.lazy(&tcx.fn_sig(def_id)),
}
} else {
bug!()
};
EntryKind::Method(self.lazy(&MethodData {
fn_data: fn_data,
container: container,
fn_data,
container,
has_self: trait_item.method_has_self_argument,
}))
}
@ -735,7 +735,7 @@ impl<'a, 'b: 'a, 'tcx: 'b> IsolatedEncoder<'a, 'b, 'tcx> {
};
Entry {
kind: kind,
kind,
visibility: self.lazy(&trait_item.vis),
span: self.lazy(&ast_item.span),
attributes: self.encode_attributes(&ast_item.attrs),
@ -805,8 +805,8 @@ impl<'a, 'b: 'a, 'tcx: 'b> IsolatedEncoder<'a, 'b, 'tcx> {
bug!()
};
EntryKind::Method(self.lazy(&MethodData {
fn_data: fn_data,
container: container,
fn_data,
container,
has_self: impl_item.method_has_self_argument,
}))
}
@ -828,7 +828,7 @@ impl<'a, 'b: 'a, 'tcx: 'b> IsolatedEncoder<'a, 'b, 'tcx> {
};
Entry {
kind: kind,
kind,
visibility: self.lazy(&impl_item.vis),
span: self.lazy(&ast_item.span),
attributes: self.encode_attributes(&ast_item.attrs),
@ -915,7 +915,7 @@ impl<'a, 'b: 'a, 'tcx: 'b> IsolatedEncoder<'a, 'b, 'tcx> {
}
hir::ItemFn(_, _, constness, .., body) => {
let data = FnData {
constness: constness,
constness,
arg_names: self.encode_fn_arg_names_for_body(body),
sig: self.lazy(&tcx.fn_sig(def_id)),
};
@ -946,7 +946,7 @@ impl<'a, 'b: 'a, 'tcx: 'b> IsolatedEncoder<'a, 'b, 'tcx> {
EntryKind::Struct(self.lazy(&VariantData {
ctor_kind: variant.ctor_kind,
discr: variant.discr,
struct_ctor: struct_ctor,
struct_ctor,
ctor_sig: None,
}), repr_options)
}
@ -998,10 +998,10 @@ impl<'a, 'b: 'a, 'tcx: 'b> IsolatedEncoder<'a, 'b, 'tcx> {
});
let data = ImplData {
polarity: polarity,
defaultness: defaultness,
polarity,
defaultness,
parent_impl: parent,
coerce_unsized_info: coerce_unsized_info,
coerce_unsized_info,
trait_ref: trait_ref.map(|trait_ref| self.lazy(&trait_ref)),
};
@ -1023,7 +1023,7 @@ impl<'a, 'b: 'a, 'tcx: 'b> IsolatedEncoder<'a, 'b, 'tcx> {
};
Entry {
kind: kind,
kind,
visibility: self.lazy(&ty::Visibility::from_hir(&item.vis, item.id, tcx)),
span: self.lazy(&item.span),
attributes: self.encode_attributes(&item.attrs),
@ -1333,7 +1333,7 @@ impl<'a, 'b: 'a, 'tcx: 'b> IsolatedEncoder<'a, 'b, 'tcx> {
debug!("IsolatedEncoder::encode_impls()");
let tcx = self.tcx;
let mut visitor = ImplVisitor {
tcx: tcx,
tcx,
impls: FxHashMap(),
};
tcx.hir.krate().visit_all_item_likes(&mut visitor);
@ -1412,7 +1412,7 @@ impl<'a, 'b: 'a, 'tcx: 'b> IsolatedEncoder<'a, 'b, 'tcx> {
};
Entry {
kind: kind,
kind,
visibility: self.lazy(&ty::Visibility::from_hir(&nitem.vis, nitem.id, tcx)),
span: self.lazy(&nitem.span),
attributes: self.encode_attributes(&nitem.attrs),
@ -1653,14 +1653,14 @@ pub fn encode_metadata<'a, 'tcx>(tcx: TyCtxt<'a, 'tcx, 'tcx>,
let (root, metadata_hashes) = {
let mut ecx = EncodeContext {
opaque: opaque::Encoder::new(&mut cursor),
tcx: tcx,
link_meta: link_meta,
exported_symbols: exported_symbols,
tcx,
link_meta,
exported_symbols,
lazy_state: LazyState::NoNode,
type_shorthands: Default::default(),
predicate_shorthands: Default::default(),
metadata_hashes: EncodedMetadataHashes::new(),
compute_ich: compute_ich,
compute_ich,
};
// Encode the rustc version string in a predictable location.

View File

@ -92,7 +92,7 @@ impl<'a, 'b, 'tcx> IndexBuilder<'a, 'b, 'tcx> {
pub fn new(ecx: &'a mut EncodeContext<'b, 'tcx>) -> Self {
IndexBuilder {
items: Index::new(ecx.tcx.hir.definitions().def_index_counts_lo_hi()),
ecx: ecx,
ecx,
}
}
@ -137,7 +137,7 @@ impl<'a, 'b, 'tcx> IndexBuilder<'a, 'b, 'tcx> {
if let Some(hash) = fingerprint {
ecx.metadata_hashes.hashes.push(EncodedMetadataHash {
def_index: id.index,
hash: hash,
hash,
});
}

View File

@ -32,8 +32,8 @@ impl<'a, 'b: 'a, 'tcx: 'b> IsolatedEncoder<'a, 'b, 'tcx> {
let tcx = ecx.tcx;
let compute_ich = ecx.compute_ich;
IsolatedEncoder {
tcx: tcx,
ecx: ecx,
tcx,
ecx,
hcx: if compute_ich {
// We are always hashing spans for things in metadata because
// don't know if a downstream crate will use them or not.

View File

@ -518,9 +518,9 @@ impl<'a> Context<'a> {
if let Some((h, m)) = slot {
libraries.insert(h,
Library {
dylib: dylib,
rlib: rlib,
rmeta: rmeta,
dylib,
rlib,
rmeta,
metadata: m,
});
}
@ -808,10 +808,10 @@ impl<'a> Context<'a> {
match slot {
Some((_, metadata)) => {
Some(Library {
dylib: dylib,
rlib: rlib,
rmeta: rmeta,
metadata: metadata,
dylib,
rlib,
rmeta,
metadata,
})
}
None => None,

View File

@ -84,7 +84,7 @@ pub struct Lazy<T> {
impl<T> Lazy<T> {
pub fn with_position(position: usize) -> Lazy<T> {
Lazy {
position: position,
position,
_marker: PhantomData,
}
}
@ -141,8 +141,8 @@ impl<T> LazySeq<T> {
pub fn with_position_and_length(position: usize, len: usize) -> LazySeq<T> {
LazySeq {
len: len,
position: position,
len,
position,
_marker: PhantomData,
}
}
@ -199,7 +199,7 @@ pub struct Tracked<T> {
impl<T> Tracked<T> {
pub fn new(state: T) -> Tracked<T> {
Tracked {
state: state,
state,
}
}

View File

@ -49,7 +49,7 @@ impl<'tcx> CFG<'tcx> {
source_info: SourceInfo,
extent: CodeExtent) {
self.push(block, Statement {
source_info: source_info,
source_info,
kind: StatementKind::EndRegion(extent),
});
}
@ -60,7 +60,7 @@ impl<'tcx> CFG<'tcx> {
lvalue: &Lvalue<'tcx>,
rvalue: Rvalue<'tcx>) {
self.push(block, Statement {
source_info: source_info,
source_info,
kind: StatementKind::Assign(lvalue.clone(), rvalue)
});
}
@ -93,8 +93,8 @@ impl<'tcx> CFG<'tcx> {
block,
self.block_data(block));
self.block_data_mut(block).terminator = Some(Terminator {
source_info: source_info,
kind: kind,
source_info,
kind,
});
}
}

View File

@ -100,7 +100,7 @@ impl<'a, 'gcx, 'tcx> Builder<'a, 'gcx, 'tcx> {
if let Some(scope) = scope {
// schedule a shallow free of that memory, lest we unwind:
this.cfg.push(block, Statement {
source_info: source_info,
source_info,
kind: StatementKind::StorageLive(result.clone())
});
this.schedule_drop(expr_span, scope, &result, value.ty);

View File

@ -52,7 +52,7 @@ impl<'a, 'gcx, 'tcx> Builder<'a, 'gcx, 'tcx> {
if !expr_ty.is_never() && temp_lifetime.is_some() {
this.cfg.push(block, Statement {
source_info: source_info,
source_info,
kind: StatementKind::StorageLive(temp.clone())
});
}

Some files were not shown because too many files have changed in this diff Show More