Fixed error starting with uppercase

Error messages cleaned in librustc/middle

Error messages cleaned in libsyntax

Error messages cleaned in libsyntax more agressively

Error messages cleaned in librustc more aggressively

Fixed affected tests

Fixed other failing tests

Last failing tests fixed
This commit is contained in:
mr.Shu 2014-02-06 10:38:08 +01:00
parent 35518514c4
commit ee3fa68fed
74 changed files with 167 additions and 167 deletions

View File

@ -76,7 +76,7 @@ pub fn WriteOutputFile(
let result = llvm::LLVMRustWriteOutputFile(
Target, PM, M, Output, FileType);
if !result {
llvm_err(sess, ~"Could not write output");
llvm_err(sess, ~"could not write output");
}
})
}
@ -189,7 +189,7 @@ pub mod write {
for pass in sess.opts.custom_passes.iter() {
pass.with_c_str(|s| {
if !llvm::LLVMRustAddPass(mpm, s) {
sess.warn(format!("Unknown pass {}, ignoring", *pass));
sess.warn(format!("unknown pass {}, ignoring", *pass));
}
})
}
@ -876,11 +876,11 @@ fn link_binary_output(sess: Session,
let obj_is_writeable = is_writeable(&obj_filename);
let out_is_writeable = is_writeable(&out_filename);
if !out_is_writeable {
sess.fatal(format!("Output file {} is not writeable -- check its permissions.",
sess.fatal(format!("output file {} is not writeable -- check its permissions.",
out_filename.display()));
}
else if !obj_is_writeable {
sess.fatal(format!("Object file {} is not writeable -- check its permissions.",
sess.fatal(format!("object file {} is not writeable -- check its permissions.",
obj_filename.display()));
}

View File

@ -308,7 +308,7 @@ impl Session_ {
// This exists to help with refactoring to eliminate impossible
// cases later on
pub fn impossible_case(&self, sp: Span, msg: &str) -> ! {
self.span_bug(sp, format!("Impossible case reached: {}", msg));
self.span_bug(sp, format!("impossible case reached: {}", msg));
}
pub fn verbose(&self) -> bool { self.debugging_opt(VERBOSE) }
pub fn time_passes(&self) -> bool { self.debugging_opt(TIME_PASSES) }

View File

@ -527,7 +527,7 @@ fn parse_sig(st: &mut PState, conv: conv_did) -> ty::FnSig {
let variadic = match next(st) {
'V' => true,
'N' => false,
r => fail!(format!("Bad variadic: {}", r)),
r => fail!(format!("bad variadic: {}", r)),
};
let ret_ty = parse_ty(st, |x,y| conv(x,y));
ty::FnSig {binder_id: id,

View File

@ -195,7 +195,7 @@ fn enc_region(w: &mut MemWriter, cx: @ctxt, r: ty::Region) {
}
ty::ReInfer(_) => {
// these should not crop up after typeck
cx.diag.handler().bug("Cannot encode region variables");
cx.diag.handler().bug("cannot encode region variables");
}
}
}
@ -320,7 +320,7 @@ fn enc_sty(w: &mut MemWriter, cx: @ctxt, st: &ty::sty) {
enc_bare_fn_ty(w, cx, f);
}
ty::ty_infer(_) => {
cx.diag.handler().bug("Cannot encode inference variable types");
cx.diag.handler().bug("cannot encode inference variable types");
}
ty::ty_param(param_ty {idx: id, def_id: did}) => {
mywrite!(w, "p{}|{}", (cx.ds)(did), id);
@ -334,7 +334,7 @@ fn enc_sty(w: &mut MemWriter, cx: @ctxt, st: &ty::sty) {
enc_substs(w, cx, substs);
mywrite!(w, "]");
}
ty::ty_err => fail!("Shouldn't encode error type")
ty::ty_err => fail!("shouldn't encode error type")
}
}

View File

@ -215,7 +215,7 @@ impl<'a> GuaranteeLifetimeContext<'a> {
// the check above should fail for anything is not ReScope
self.bccx.tcx.sess.span_bug(
cmt_base.span,
format!("Cannot issue root for scope region: {:?}",
format!("cannot issue root for scope region: {:?}",
self.loan_region));
}
};

View File

@ -496,7 +496,7 @@ impl<'a> GatherLoanCtxt<'a> {
ty::ReInfer(..) => {
self.tcx().sess.span_bug(
cmt.span,
format!("Invalid borrow lifetime: {:?}", loan_region));
format!("invalid borrow lifetime: {:?}", loan_region));
}
};
debug!("loan_scope = {:?}", loan_scope);
@ -820,7 +820,7 @@ impl<'a> GatherLoanCtxt<'a> {
_ => {
self.tcx().sess.span_bug(
pat.span,
format!("Type of slice pattern is not a slice"));
format!("type of slice pattern is not a slice"));
}
}
}

View File

@ -769,14 +769,14 @@ impl BorrowckCtxt {
}
_ => {
self.tcx.sess.bug(
format!("Loan path LpVar({:?}) maps to {:?}, not local",
format!("loan path LpVar({:?}) maps to {:?}, not local",
id, pat));
}
}
}
r => {
self.tcx.sess.bug(
format!("Loan path LpVar({:?}) maps to {:?}, not local",
format!("loan path LpVar({:?}) maps to {:?}, not local",
id, r));
}
}

View File

@ -494,7 +494,7 @@ impl MoveData {
dfcx_assign.add_kill(kill_id, assignment_index);
}
LpExtend(..) => {
tcx.sess.bug("Var assignment for non var path");
tcx.sess.bug("var assignment for non var path");
}
}
}

View File

@ -507,13 +507,13 @@ impl CFGBuilder {
}
self.tcx.sess.span_bug(
expr.span,
format!("No loop scope for id {:?}", loop_id));
format!("no loop scope for id {:?}", loop_id));
}
r => {
self.tcx.sess.span_bug(
expr.span,
format!("Bad entry `{:?}` in def_map for label", r));
format!("bad entry `{:?}` in def_map for label", r));
}
}
}

View File

@ -971,7 +971,7 @@ fn check_legality_of_move_bindings(cx: &MatchCheckCtxt,
_ => {
cx.tcx.sess.span_bug(
p.span,
format!("Binding pattern {} is \
format!("binding pattern {} is \
not an identifier: {:?}",
p.id, p.node));
}

View File

@ -342,8 +342,8 @@ pub fn eval_const_expr_partial<T: ty::ExprTyProvider>(tcx: &T, e: &Expr)
Ok(const_float(f)) => Ok(const_float(-f)),
Ok(const_int(i)) => Ok(const_int(-i)),
Ok(const_uint(i)) => Ok(const_uint(-i)),
Ok(const_str(_)) => Err(~"Negate on string"),
Ok(const_bool(_)) => Err(~"Negate on boolean"),
Ok(const_str(_)) => Err(~"negate on string"),
Ok(const_bool(_)) => Err(~"negate on boolean"),
ref err => ((*err).clone())
}
}
@ -352,7 +352,7 @@ pub fn eval_const_expr_partial<T: ty::ExprTyProvider>(tcx: &T, e: &Expr)
Ok(const_int(i)) => Ok(const_int(!i)),
Ok(const_uint(i)) => Ok(const_uint(!i)),
Ok(const_bool(b)) => Ok(const_bool(!b)),
_ => Err(~"Not on float or string")
_ => Err(~"not on float or string")
}
}
ExprBinary(_, op, a, b) => {
@ -371,7 +371,7 @@ pub fn eval_const_expr_partial<T: ty::ExprTyProvider>(tcx: &T, e: &Expr)
BiNe => fromb(a != b),
BiGe => fromb(a >= b),
BiGt => fromb(a > b),
_ => Err(~"Can't do this op on floats")
_ => Err(~"can't do this op on floats")
}
}
(Ok(const_int(a)), Ok(const_int(b))) => {
@ -423,14 +423,14 @@ pub fn eval_const_expr_partial<T: ty::ExprTyProvider>(tcx: &T, e: &Expr)
match op {
BiShl => Ok(const_int(a << b)),
BiShr => Ok(const_int(a >> b)),
_ => Err(~"Can't do this op on an int and uint")
_ => Err(~"can't do this op on an int and uint")
}
}
(Ok(const_uint(a)), Ok(const_int(b))) => {
match op {
BiShl => Ok(const_uint(a << b)),
BiShr => Ok(const_uint(a >> b)),
_ => Err(~"Can't do this op on a uint and int")
_ => Err(~"can't do this op on a uint and int")
}
}
(Ok(const_bool(a)), Ok(const_bool(b))) => {
@ -442,10 +442,10 @@ pub fn eval_const_expr_partial<T: ty::ExprTyProvider>(tcx: &T, e: &Expr)
BiBitOr => a | b,
BiEq => a == b,
BiNe => a != b,
_ => return Err(~"Can't do this op on bools")
_ => return Err(~"can't do this op on bools")
}))
}
_ => Err(~"Bad operands for binary")
_ => Err(~"bad operands for binary")
}
}
ExprCast(base, target_ty) => {
@ -456,7 +456,7 @@ pub fn eval_const_expr_partial<T: ty::ExprTyProvider>(tcx: &T, e: &Expr)
.or_else(|| astconv::ast_ty_to_prim_ty(tcx.ty_ctxt(), target_ty))
.unwrap_or_else(|| tcx.ty_ctxt().sess.span_fatal(
target_ty.span,
format!("Target type not found for const cast")
format!("target type not found for const cast")
));
let base = eval_const_expr_partial(tcx, base);
@ -469,7 +469,7 @@ pub fn eval_const_expr_partial<T: ty::ExprTyProvider>(tcx: &T, e: &Expr)
const_uint(u) => Ok(const_float(u as f64)),
const_int(i) => Ok(const_float(i as f64)),
const_float(f) => Ok(const_float(f)),
_ => Err(~"Can't cast float to str"),
_ => Err(~"can't cast float to str"),
}
}
ty::ty_uint(_) => {
@ -477,7 +477,7 @@ pub fn eval_const_expr_partial<T: ty::ExprTyProvider>(tcx: &T, e: &Expr)
const_uint(u) => Ok(const_uint(u)),
const_int(i) => Ok(const_uint(i as u64)),
const_float(f) => Ok(const_uint(f as u64)),
_ => Err(~"Can't cast str to uint"),
_ => Err(~"can't cast str to uint"),
}
}
ty::ty_int(_) | ty::ty_bool => {
@ -485,10 +485,10 @@ pub fn eval_const_expr_partial<T: ty::ExprTyProvider>(tcx: &T, e: &Expr)
const_uint(u) => Ok(const_int(u as i64)),
const_int(i) => Ok(const_int(i)),
const_float(f) => Ok(const_int(f as i64)),
_ => Err(~"Can't cast str to int"),
_ => Err(~"can't cast str to int"),
}
}
_ => Err(~"Can't cast this type")
_ => Err(~"can't cast this type")
}
}
}
@ -496,14 +496,14 @@ pub fn eval_const_expr_partial<T: ty::ExprTyProvider>(tcx: &T, e: &Expr)
ExprPath(_) => {
match lookup_const(tcx.ty_ctxt(), e) {
Some(actual_e) => eval_const_expr_partial(&tcx.ty_ctxt(), actual_e),
None => Err(~"Non-constant path in constant expr")
None => Err(~"non-constant path in constant expr")
}
}
ExprLit(lit) => Ok(lit_to_const(lit)),
// If we have a vstore, just keep going; it has to be a string
ExprVstore(e, _) => eval_const_expr_partial(tcx, e),
ExprParen(e) => eval_const_expr_partial(tcx, e),
_ => Err(~"Unsupported constant expr")
_ => Err(~"unsupported constant expr")
}
}

View File

@ -879,7 +879,7 @@ impl<'a, O:DataFlowOperator> PropagationContext<'a, O> {
None => {
self.tcx().sess.span_bug(
expr.span,
format!("No loop scope for id {:?}", loop_id));
format!("no loop scope for id {:?}", loop_id));
}
}
}
@ -887,7 +887,7 @@ impl<'a, O:DataFlowOperator> PropagationContext<'a, O> {
r => {
self.tcx().sess.span_bug(
expr.span,
format!("Bad entry `{:?}` in def_map for label", r));
format!("bad entry `{:?}` in def_map for label", r));
}
}
}

View File

@ -328,7 +328,7 @@ impl IrMaps {
Some(&var) => var,
None => {
self.tcx.sess.span_bug(
span, format!("No variable registered for id {}", node_id));
span, format!("no variable registered for id {}", node_id));
}
}
}
@ -627,7 +627,7 @@ impl Liveness {
// code have to agree about which AST nodes are worth
// creating liveness nodes for.
self.tcx.sess.span_bug(
span, format!("No live node registered for node {}",
span, format!("no live node registered for node {}",
node_id));
}
}
@ -759,7 +759,7 @@ impl Liveness {
let def_map = self.tcx.def_map.borrow();
match def_map.get().find(&id) {
Some(&DefLabel(loop_id)) => loop_id,
_ => self.tcx.sess.span_bug(sp, "Label on break/loop \
_ => self.tcx.sess.span_bug(sp, "label on break/loop \
doesn't refer to a loop")
}
}
@ -1152,7 +1152,7 @@ impl Liveness {
match break_ln.get().find(&sc) {
Some(&b) => b,
None => self.tcx.sess.span_bug(expr.span,
"Break to unknown label")
"break to unknown label")
}
}
@ -1167,7 +1167,7 @@ impl Liveness {
match cont_ln.get().find(&sc) {
Some(&b) => b,
None => self.tcx.sess.span_bug(expr.span,
"Loop to unknown label")
"loop to unknown label")
}
}

View File

@ -548,7 +548,7 @@ impl mem_categorization_ctxt {
_ => {
self.tcx.sess.span_bug(
span,
format!("Upvar of non-closure {:?} - {}",
format!("upvar of non-closure {:?} - {}",
fn_node_id, ty.repr(self.tcx)));
}
}
@ -652,7 +652,7 @@ impl mem_categorization_ctxt {
None => {
self.tcx.sess.span_bug(
node.span(),
format!("Explicit deref of non-derefable type: {}",
format!("explicit deref of non-derefable type: {}",
ty_to_str(self.tcx, base_cmt.ty)));
}
};
@ -745,7 +745,7 @@ impl mem_categorization_ctxt {
None => {
self.tcx.sess.span_bug(
elt.span(),
format!("Explicit index of non-index type `{}`",
format!("explicit index of non-index type `{}`",
ty_to_str(self.tcx, base_cmt.ty)));
}
};

View File

@ -142,7 +142,7 @@ impl RegionMaps {
let scope_map = self.scope_map.borrow();
match scope_map.get().find(&id) {
Some(&r) => r,
None => { fail!("No enclosing scope for id {}", id); }
None => { fail!("no enclosing scope for id {}", id); }
}
}
@ -154,7 +154,7 @@ impl RegionMaps {
let var_map = self.var_map.borrow();
match var_map.get().find(&var_id) {
Some(&r) => r,
None => { fail!("No enclosing scope for id {}", var_id); }
None => { fail!("no enclosing scope for id {}", var_id); }
}
}
@ -554,7 +554,7 @@ fn resolve_local(visitor: &mut RegionResolutionVisitor,
None => {
visitor.sess.span_bug(
local.span,
"Local without enclosing block");
"local without enclosing block");
}
};

View File

@ -1378,7 +1378,7 @@ impl Resolver {
}
match method_map.get().find_mut(name) {
Some(s) => { s.insert(def_id); },
_ => fail!("Can't happen"),
_ => fail!("can't happen"),
}
}
@ -1696,7 +1696,7 @@ impl Resolver {
}
match method_map.get().find_mut(name) {
Some(s) => { s.insert(def_id); },
_ => fail!("Can't happen"),
_ => fail!("can't happen"),
}
}
@ -3992,7 +3992,7 @@ impl Resolver {
self.resolve_error(field.span,
format!("field `{}` is already declared", ident_str));
self.session.span_note(prev_field.span,
"Previously declared here");
"previously declared here");
},
None => {
ident_map.insert(ident, field);
@ -4476,7 +4476,7 @@ impl Resolver {
// in the same disjunct, which is an
// error
self.resolve_error(pattern.span,
format!("Identifier `{}` is bound more \
format!("identifier `{}` is bound more \
than once in the same pattern",
path_to_str(path, self.session
.intr())));

View File

@ -297,7 +297,7 @@ impl LifetimeContext {
def: ast::DefRegion) {
if lifetime_ref.id == ast::DUMMY_NODE_ID {
self.sess.span_bug(lifetime_ref.span,
"Lifetime reference not renumbered, \
"lifetime reference not renumbered, \
probably a bug in syntax::fold");
}

View File

@ -484,7 +484,7 @@ fn assert_is_binding_or_wild(bcx: &Block, p: @ast::Pat) {
if !pat_is_binding_or_wild(bcx.tcx().def_map, p) {
bcx.sess().span_bug(
p.span,
format!("Expected an identifier pattern but found p: {}",
format!("expected an identifier pattern but found p: {}",
p.repr(bcx.tcx())));
}
}

View File

@ -434,7 +434,7 @@ fn generic_type_of(cx: &CrateContext, r: &Repr, name: Option<&str>, sizing: bool
Type::array(&Type::i64(), align_units),
a if a.population_count() == 1 => Type::array(&Type::vector(&Type::i32(), a / 4),
align_units),
_ => fail!("Unsupported enum alignment: {:?}", align)
_ => fail!("unsupported enum alignment: {:?}", align)
};
assert_eq!(machine::llalign_of_min(cx, pad_ty) as u64, align);
assert_eq!(align % discr_size, 0);

View File

@ -103,7 +103,7 @@ fn trans<'a>(bcx: &'a Block<'a>, expr: &ast::Expr) -> Callee<'a> {
_ => {
bcx.tcx().sess.span_bug(
expr.span,
format!("Type of callee is neither bare-fn nor closure: {}",
format!("type of callee is neither bare-fn nor closure: {}",
bcx.ty_to_str(datum.ty)));
}
}
@ -151,7 +151,7 @@ fn trans<'a>(bcx: &'a Block<'a>, expr: &ast::Expr) -> Callee<'a> {
ast::DefSelfTy(..) | ast::DefMethod(..) => {
bcx.tcx().sess.span_bug(
ref_expr.span,
format!("Cannot translate def {:?} \
format!("cannot translate def {:?} \
to a callable thing!", def));
}
}

View File

@ -211,7 +211,7 @@ impl<'a> CleanupMethods<'a> for FunctionContext<'a> {
_ => {}
}
}
self.ccx.tcx.sess.bug("No loop scope found");
self.ccx.tcx.sess.bug("no loop scope found");
}
fn normal_exit_block(&'a self,
@ -337,7 +337,7 @@ impl<'a> CleanupMethods<'a> for FunctionContext<'a> {
}
self.ccx.tcx.sess.bug(
format!("No cleanup scope {} found",
format!("no cleanup scope {} found",
ast_map::node_id_to_str(self.ccx.tcx.items, cleanup_scope,
token::get_ident_interner())));
}
@ -548,7 +548,7 @@ impl<'a> CleanupHelperMethods<'a> for FunctionContext<'a> {
LoopExit(id, _) => {
self.ccx.tcx.sess.bug(format!(
"Cannot exit from scope {:?}, \
"cannot exit from scope {:?}, \
not in scope", id));
}
}

View File

@ -472,7 +472,7 @@ impl<'a> Block<'a> {
Some(&v) => v,
None => {
self.tcx().sess.bug(format!(
"No def associated with node id {:?}", nid));
"no def associated with node id {:?}", nid));
}
}
}
@ -852,7 +852,7 @@ pub fn node_id_type_params(bcx: &Block, id: ast::NodeId) -> ~[ty::t] {
if !params.iter().all(|t| !ty::type_needs_infer(*t)) {
bcx.sess().bug(
format!("Type parameters for node {} include inference types: {}",
format!("type parameters for node {} include inference types: {}",
id, params.map(|t| bcx.ty_to_str(*t)).connect(",")));
}

View File

@ -146,14 +146,14 @@ fn const_deref(cx: &CrateContext, v: ValueRef, t: ty::t, explicit: bool)
const_deref_newtype(cx, v, t)
}
_ => {
cx.sess.bug(format!("Unexpected dereferenceable type {}",
cx.sess.bug(format!("unexpected dereferenceable type {}",
ty_to_str(cx.tcx, t)))
}
};
(dv, mt.ty)
}
None => {
cx.sess.bug(format!("Can't dereference const of type {}",
cx.sess.bug(format!("can't dereference const of type {}",
ty_to_str(cx.tcx, t)))
}
}

View File

@ -298,7 +298,7 @@ pub fn create_local_var_metadata(bcx: &Block, local: &ast::Local) {
Some(datum) => datum,
None => {
bcx.tcx().sess.span_bug(span,
format!("No entry in lllocals table for {:?}",
format!("no entry in lllocals table for {:?}",
node_id));
}
}
@ -440,7 +440,7 @@ pub fn create_argument_metadata(bcx: &Block, arg: &ast::Arg) {
Some(v) => v,
None => {
bcx.tcx().sess.span_bug(span,
format!("No entry in llargs table for {:?}",
format!("no entry in llargs table for {:?}",
node_id));
}
}

View File

@ -81,17 +81,17 @@ pub fn llvm_calling_convention(ccx: &CrateContext,
match abi {
RustIntrinsic => {
// Intrinsics are emitted by monomorphic fn
ccx.sess.bug(format!("Asked to register intrinsic fn"));
ccx.sess.bug(format!("asked to register intrinsic fn"));
}
Rust => {
// FIXME(#3678) Implement linking to foreign fns with Rust ABI
ccx.sess.unimpl(
format!("Foreign functions with Rust ABI"));
format!("foreign functions with Rust ABI"));
}
// It's the ABI's job to select this, not us.
System => ccx.sess.bug("System abi should be selected elsewhere"),
System => ccx.sess.bug("system abi should be selected elsewhere"),
Stdcall => lib::llvm::X86StdcallCallConv,
Fastcall => lib::llvm::X86FastcallCallConv,
@ -365,7 +365,7 @@ pub fn trans_foreign_mod(ccx: @CrateContext,
(abis, (*path).clone())
}
_ => {
fail!("Unable to find foreign item in tcx.items \
fail!("unable to find foreign item in tcx.items \
table.")
}
};

View File

@ -227,7 +227,7 @@ pub fn trans_intrinsic(ccx: @CrateContext,
"acq" => lib::llvm::Acquire,
"rel" => lib::llvm::Release,
"acqrel" => lib::llvm::AcquireRelease,
_ => ccx.sess.fatal("Unknown ordering in atomic intrinsic")
_ => ccx.sess.fatal("unknown ordering in atomic intrinsic")
}
};
@ -268,7 +268,7 @@ pub fn trans_intrinsic(ccx: @CrateContext,
"min" => lib::llvm::Min,
"umax" => lib::llvm::UMax,
"umin" => lib::llvm::UMin,
_ => ccx.sess.fatal("Unknown atomic operation")
_ => ccx.sess.fatal("unknown atomic operation")
};
let old = AtomicRMW(bcx, atom_op, get_param(decl, first_real_arg),

View File

@ -98,7 +98,7 @@ pub fn monomorphic_fn(ccx: @CrateContext,
session::expect(
ccx.sess,
ccx.tcx.items.find(fn_id.node),
|| format!("While monomorphizing {:?}, couldn't find it in the \
|| format!("while monomorphizing {:?}, couldn't find it in the \
item map (may have attempted to monomorphize an item \
defined in a different crate?)", fn_id))
};
@ -286,7 +286,7 @@ pub fn monomorphic_fn(ccx: @CrateContext,
d
}
_ => {
ccx.tcx.sess.bug(format!("Can't monomorphize a {:?}",
ccx.tcx.sess.bug(format!("can't monomorphize a {:?}",
map_node))
}
}
@ -310,7 +310,7 @@ pub fn monomorphic_fn(ccx: @CrateContext,
ast_map::NodeBlock(..) |
ast_map::NodeCalleeScope(..) |
ast_map::NodeLocal(..) => {
ccx.tcx.sess.bug(format!("Can't monomorphize a {:?}", map_node))
ccx.tcx.sess.bug(format!("can't monomorphize a {:?}", map_node))
}
};

View File

@ -95,7 +95,7 @@ impl<'a> Reflector<'a> {
let tcx = self.bcx.tcx();
let mth_idx = ty::method_idx(
tcx.sess.ident_of(~"visit_" + ty_name),
*self.visitor_methods).expect(format!("Couldn't find visit method \
*self.visitor_methods).expect(format!("couldn't find visit method \
for {}", ty_name));
let mth_ty =
ty::mk_bare_fn(tcx, self.visitor_methods[mth_idx].fty.clone());

View File

@ -404,7 +404,7 @@ pub fn write_content<'a>(
}
_ => {
bcx.tcx().sess.span_bug(content_expr.span,
"Unexpected evec content");
"unexpected evec content");
}
}
}
@ -465,7 +465,7 @@ pub fn write_content<'a>(
}
_ => {
bcx.tcx().sess.span_bug(content_expr.span,
"Unexpected vec content");
"unexpected vec content");
}
}
}
@ -498,7 +498,7 @@ pub fn elements_required(bcx: &Block, content_expr: &ast::Expr) -> uint {
ast::LitStr(ref s, _) => s.get().len(),
_ => {
bcx.tcx().sess.span_bug(content_expr.span,
"Unexpected evec content")
"unexpected evec content")
}
}
},
@ -507,7 +507,7 @@ pub fn elements_required(bcx: &Block, content_expr: &ast::Expr) -> uint {
ty::eval_repeat_count(&bcx.tcx(), count_expr)
}
_ => bcx.tcx().sess.span_bug(content_expr.span,
"Unexpected vec content")
"unexpected vec content")
}
}

View File

@ -2101,7 +2101,7 @@ pub fn type_contents(cx: ctxt, ty: t) -> TypeContents {
ty_type => TC::None,
ty_err => {
cx.sess.bug("Asked to compute contents of error type");
cx.sess.bug("asked to compute contents of error type");
}
};
@ -2926,7 +2926,7 @@ pub fn adjust_ty(cx: ctxt,
None => {
cx.sess.span_bug(
span,
format!("The {}th autoderef failed: \
format!("the {}th autoderef failed: \
{}",
i,
ty_to_str(cx, adjusted_ty)));
@ -3132,7 +3132,7 @@ pub fn resolve_expr(tcx: ctxt, expr: &ast::Expr) -> ast::Def {
Some(&def) => def,
None => {
tcx.sess.span_bug(expr.span, format!(
"No def-map entry for expr {:?}", expr.id));
"no def-map entry for expr {:?}", expr.id));
}
}
}
@ -3209,7 +3209,7 @@ pub fn expr_kind(tcx: ctxt,
def => {
tcx.sess.span_bug(expr.span, format!(
"Uncategorized def for expr {:?}: {:?}",
"uncategorized def for expr {:?}: {:?}",
expr.id, def));
}
}
@ -3335,7 +3335,7 @@ pub fn field_idx_strict(tcx: ty::ctxt, name: ast::Name, fields: &[field])
for f in fields.iter() { if f.ident.name == name { return i; } i += 1u; }
let string = token::get_ident(name);
tcx.sess.bug(format!(
"No field named `{}` found in the list of fields `{:?}`",
"no field named `{}` found in the list of fields `{:?}`",
string.get(),
fields.map(|f| tcx.sess.str_of(f.ident))));
}
@ -3687,7 +3687,7 @@ fn lookup_locally_or_in_crate_store<V:Clone>(
}
if def_id.crate == ast::LOCAL_CRATE {
fail!("No def'n found for {:?} in tcx.{}", def_id, descr);
fail!("no def'n found for {:?} in tcx.{}", def_id, descr);
}
let v = load_external();
map.insert(def_id, v.clone());

View File

@ -789,7 +789,7 @@ impl<'a> LookupContext<'a> {
ty_err => None,
ty_unboxed_vec(_) | ty_type | ty_infer(TyVar(_)) => {
self.bug(format!("Unexpected type: {}",
self.bug(format!("unexpected type: {}",
self.ty_to_str(self_ty)));
}
}

View File

@ -517,9 +517,9 @@ pub fn check_no_duplicate_fields(tcx: ty::ctxt,
let orig_sp = field_names.find(&id).map(|x| *x);
match orig_sp {
Some(orig_sp) => {
tcx.sess.span_err(sp, format!("Duplicate field name {} in record type declaration",
tcx.sess.span_err(sp, format!("duplicate field name {} in record type declaration",
tcx.sess.str_of(id)));
tcx.sess.span_note(orig_sp, "First declaration of this field occurred here");
tcx.sess.span_note(orig_sp, "first declaration of this field occurred here");
break;
}
None => {
@ -1006,7 +1006,7 @@ impl FnCtxt {
None => {
self.tcx().sess.span_bug(
span,
format!("No type for local variable {:?}", nid));
format!("no type for local variable {:?}", nid));
}
}
}
@ -1619,7 +1619,7 @@ pub fn check_expr_with_unifier(fcx: @FnCtxt,
_ => {
fcx.tcx().sess.span_bug(
sp,
format!("Method without bare fn type"));
format!("method without bare fn type"));
}
}
}

View File

@ -473,7 +473,7 @@ fn check_expr_fn_block(rcx: &mut Rcx,
_ => {
tcx.sess.span_bug(
expr.span,
"Expected expr_fn_block");
"expected expr_fn_block");
}
}
}
@ -1189,7 +1189,7 @@ pub mod guarantor {
None => {
tcx.sess.span_bug(
expr.span,
format!("Autoderef but type not derefable: {}",
format!("autoderef but type not derefable: {}",
ty_to_str(tcx, ct.ty)));
}
}

View File

@ -430,7 +430,7 @@ pub fn ensure_supertraits(ccx: &CrateCtxt,
if ty_trait_refs.iter().any(|other_trait| other_trait.def_id == trait_ref.def_id) {
// This means a trait inherited from the same supertrait more
// than once.
tcx.sess.span_err(sp, "Duplicate supertrait in trait declaration");
tcx.sess.span_err(sp, "duplicate supertrait in trait declaration");
break;
} else {
ty_trait_refs.push(trait_ref);
@ -891,7 +891,7 @@ pub fn ty_of_item(ccx: &CrateCtxt, it: &ast::Item)
ast::ItemTrait(..) => {
tcx.sess.span_bug(
it.span,
format!("Invoked ty_of_item on trait"));
format!("invoked ty_of_item on trait"));
}
ast::ItemStruct(_, ref generics) => {
let ty_generics = ty_generics(ccx, generics, 0);

View File

@ -202,7 +202,7 @@ impl<'f> Coerce<'f> {
Err(e) => {
self.get_ref().infcx.tcx.sess.span_bug(
self.get_ref().trace.origin.span(),
format!("Failed to resolve even without \
format!("failed to resolve even without \
any force options: {:?}", e));
}
}

View File

@ -528,7 +528,7 @@ pub fn var_ids<T:Combine>(this: &T,
r => {
this.infcx().tcx.sess.span_bug(
this.trace().origin.span(),
format!("Found non-region-vid: {:?}", r));
format!("found non-region-vid: {:?}", r));
}
}).collect()
}

View File

@ -233,7 +233,7 @@ impl RegionVarBindings {
self.bound_count.set(sc + 1);
if sc >= self.bound_count.get() {
self.tcx.sess.bug("Rollover in RegionInference new_bound()");
self.tcx.sess.bug("rollover in RegionInference new_bound()");
}
ReLateBound(binder_id, BrFresh(sc))
@ -278,7 +278,7 @@ impl RegionVarBindings {
(_, ReLateBound(..)) => {
self.tcx.sess.span_bug(
origin.span(),
format!("Cannot relate bound region: {} <= {}",
format!("cannot relate bound region: {} <= {}",
sub.repr(self.tcx),
sup.repr(self.tcx)));
}
@ -351,7 +351,7 @@ impl RegionVarBindings {
let var_origins = self.var_origins.borrow();
self.tcx.sess.span_bug(
var_origins.get()[rid.to_uint()].span(),
format!("Attempt to resolve region variable before \
format!("attempt to resolve region variable before \
values have been computed!"))
}
Some(ref values) => values[rid.to_uint()]
@ -544,7 +544,7 @@ impl RegionVarBindings {
(ReEarlyBound(..), _) |
(_, ReEarlyBound(..)) => {
self.tcx.sess.bug(
format!("Cannot relate bound region: LUB({}, {})",
format!("cannot relate bound region: LUB({}, {})",
a.repr(self.tcx),
b.repr(self.tcx)));
}
@ -646,7 +646,7 @@ impl RegionVarBindings {
(ReEarlyBound(..), _) |
(_, ReEarlyBound(..)) => {
self.tcx.sess.bug(
format!("Cannot relate bound region: GLB({}, {})",
format!("cannot relate bound region: GLB({}, {})",
a.repr(self.tcx),
b.repr(self.tcx)));
}

View File

@ -97,7 +97,7 @@ impl Env {
return match search_mod(self, &self.crate.node.module, 0, names) {
Some(id) => id,
None => {
fail!("No item found: `%s`", names.connect("::"));
fail!("no item found: `%s`", names.connect("::"));
}
};
@ -230,7 +230,7 @@ impl Env {
for msg in self.err_messages.iter() {
debug!("Error encountered: %s", *msg);
}
format!("Resolving regions encountered %u errors but expected %u!",
format!("resolving regions encountered %u errors but expected %u!",
self.err_messages.len(),
exp_count);
}
@ -240,7 +240,7 @@ impl Env {
pub fn check_lub(&self, t1: ty::t, t2: ty::t, t_lub: ty::t) {
match self.lub().tys(t1, t2) {
Err(e) => {
fail!("Unexpected error computing LUB: %?", e)
fail!("unexpected error computing LUB: %?", e)
}
Ok(t) => {
self.assert_eq(t, t_lub);
@ -262,7 +262,7 @@ impl Env {
self.ty_to_str(t_glb));
match self.glb().tys(t1, t2) {
Err(e) => {
fail!("Unexpected error computing LUB: %?", e)
fail!("unexpected error computing LUB: %?", e)
}
Ok(t) => {
self.assert_eq(t, t_glb);
@ -281,7 +281,7 @@ impl Env {
match self.lub().tys(t1, t2) {
Err(_) => {}
Ok(t) => {
fail!("Unexpected success computing LUB: %?", self.ty_to_str(t))
fail!("unexpected success computing LUB: %?", self.ty_to_str(t))
}
}
}
@ -291,7 +291,7 @@ impl Env {
match self.glb().tys(t1, t2) {
Err(_) => {}
Ok(t) => {
fail!("Unexpected success computing GLB: %?", self.ty_to_str(t))
fail!("unexpected success computing GLB: %?", self.ty_to_str(t))
}
}
}

View File

@ -718,7 +718,7 @@ impl<'a> ConstraintContext<'a> {
ty::ty_infer(..) | ty::ty_err |
ty::ty_type | ty::ty_unboxed_vec(..) => {
self.tcx().sess.bug(
format!("Unexpected type encountered in \
format!("unexpected type encountered in \
variance inference: {}",
ty.repr(self.tcx())));
}
@ -804,7 +804,7 @@ impl<'a> ConstraintContext<'a> {
ty::ReEmpty => {
// We don't expect to see anything but 'static or bound
// regions when visiting member types or method types.
self.tcx().sess.bug(format!("Unexpected region encountered in \
self.tcx().sess.bug(format!("unexpected region encountered in \
variance inference: {}",
region.repr(self.tcx())));
}

View File

@ -64,12 +64,12 @@ fn add_bytes_to_bits<T: Int + CheckedAdd + ToBits>(bits: T, bytes: T) -> T {
let (new_high_bits, new_low_bits) = bytes.to_bits();
if new_high_bits > Zero::zero() {
fail!("Numeric overflow occured.")
fail!("numeric overflow occured.")
}
match bits.checked_add(&new_low_bits) {
Some(x) => return x,
None => fail!("Numeric overflow occured.")
None => fail!("numeric overflow occured.")
}
}

View File

@ -323,7 +323,7 @@ fn cannot_combine(n: Abi, m: Abi) {
(m == a && n == b));
}
None => {
fail!("Invalid match not detected");
fail!("invalid match not detected");
}
}
}
@ -335,7 +335,7 @@ fn can_combine(n: Abi, m: Abi) {
set.add(m);
match set.check_valid() {
Some((_, _)) => {
fail!("Valid match declared invalid");
fail!("valid match declared invalid");
}
None => {}
}

View File

@ -40,7 +40,7 @@ pub fn expand_syntax_ext(cx: &mut ExtCtxt, sp: Span, tts: &[ast::TokenTree]) ->
// u8 literal, push to vector expression
ast::LitUint(v, ast::TyU8) => {
if v > 0xFF {
cx.span_err(expr.span, "Too large u8 literal in bytes!")
cx.span_err(expr.span, "too large u8 literal in bytes!")
} else {
bytes.push(cx.expr_u8(expr.span, v as u8));
}
@ -49,9 +49,9 @@ pub fn expand_syntax_ext(cx: &mut ExtCtxt, sp: Span, tts: &[ast::TokenTree]) ->
// integer literal, push to vector expression
ast::LitIntUnsuffixed(v) => {
if v > 0xFF {
cx.span_err(expr.span, "Too large integer literal in bytes!")
cx.span_err(expr.span, "too large integer literal in bytes!")
} else if v < 0 {
cx.span_err(expr.span, "Negative integer literal in bytes!")
cx.span_err(expr.span, "negative integer literal in bytes!")
} else {
bytes.push(cx.expr_u8(expr.span, v as u8));
}
@ -62,14 +62,14 @@ pub fn expand_syntax_ext(cx: &mut ExtCtxt, sp: Span, tts: &[ast::TokenTree]) ->
if char::from_u32(v).unwrap().is_ascii() {
bytes.push(cx.expr_u8(expr.span, v as u8));
} else {
cx.span_err(expr.span, "Non-ascii char literal in bytes!")
cx.span_err(expr.span, "non-ascii char literal in bytes!")
}
}
_ => cx.span_err(expr.span, "Unsupported literal in bytes!")
_ => cx.span_err(expr.span, "unsupported literal in bytes!")
},
_ => cx.span_err(expr.span, "Non-literal in bytes!")
_ => cx.span_err(expr.span, "non-literal in bytes!")
}
}

View File

@ -92,10 +92,10 @@ fn cs_clone(
all_fields = af;
},
EnumNonMatching(..) => cx.span_bug(trait_span,
format!("Non-matching enum variants in `deriving({})`",
format!("non-matching enum variants in `deriving({})`",
name)),
StaticEnum(..) | StaticStruct(..) => cx.span_bug(trait_span,
format!("Static method in `deriving({})`",
format!("static method in `deriving({})`",
name))
}

View File

@ -75,7 +75,7 @@ fn cs_op(less: bool, equal: bool, cx: &mut ExtCtxt, span: Span, substr: &Substru
*/
let other_f = match other_fs {
[o_f] => o_f,
_ => cx.span_bug(span, "Not exactly 2 arguments in `deriving(Ord)`")
_ => cx.span_bug(span, "not exactly 2 arguments in `deriving(Ord)`")
};
let cmp = cx.expr_binary(span, op, self_f, other_f);
@ -99,7 +99,7 @@ fn cs_op(less: bool, equal: bool, cx: &mut ExtCtxt, span: Span, substr: &Substru
} else {
self_var > other_var
}),
_ => cx.span_bug(span, "Not exactly 2 arguments in `deriving(Ord)`")
_ => cx.span_bug(span, "not exactly 2 arguments in `deriving(Ord)`")
}
},
cx, span, substr)

View File

@ -110,7 +110,7 @@ pub fn cs_cmp(cx: &mut ExtCtxt, span: Span,
let order = ordering_const(cx, span, self_var.cmp(&other_var));
cx.expr_path(order)
}
_ => cx.span_bug(span, "Not exactly 2 arguments in `deriving(TotalOrd)`")
_ => cx.span_bug(span, "not exactly 2 arguments in `deriving(TotalOrd)`")
}
},
cx, span, substr)

View File

@ -656,7 +656,7 @@ impl<'a> MethodDef<'a> {
}).collect()
}
[] => { trait_.cx.span_bug(trait_.span,
"No self arguments to non-static method \
"no self arguments to non-static method \
in generic `deriving`") }
};
@ -840,7 +840,7 @@ impl<'a> MethodDef<'a> {
let index = match matching {
Some(i) => i,
None => cx.span_bug(trait_.span,
"Non-matching variants when required to \
"non-matching variants when required to \
be matching in generic `deriving`")
};
@ -965,7 +965,7 @@ impl<'a> TraitDef<'a> {
match (just_spans.is_empty(), named_idents.is_empty()) {
(false, false) => self.cx.span_bug(self.span,
"A struct with named and unnamed \
"a struct with named and unnamed \
fields in generic `deriving`"),
// named fields
(_, false) => Named(named_idents),
@ -1019,7 +1019,7 @@ impl<'a> TraitDef<'a> {
None
}
_ => {
cx.span_bug(sp, "A struct with named and unnamed fields in `deriving`");
cx.span_bug(sp, "a struct with named and unnamed fields in `deriving`");
}
};
let path = cx.path_ident(sp, cx.ident_of(format!("{}_{}", prefix, i)));
@ -1116,7 +1116,7 @@ pub fn cs_fold(use_foldl: bool,
*all_enums,
substructure.nonself_args),
StaticEnum(..) | StaticStruct(..) => {
cx.span_bug(trait_span, "Static function in `deriving`")
cx.span_bug(trait_span, "static function in `deriving`")
}
}
}
@ -1154,7 +1154,7 @@ pub fn cs_same_method(f: |&mut ExtCtxt, Span, ~[@Expr]| -> @Expr,
*all_enums,
substructure.nonself_args),
StaticEnum(..) | StaticStruct(..) => {
cx.span_bug(trait_span, "Static function in `deriving`")
cx.span_bug(trait_span, "static function in `deriving`")
}
}
}

View File

@ -48,7 +48,7 @@ pub fn expand_deriving_iter_bytes(cx: &mut ExtCtxt,
fn iter_bytes_substructure(cx: &mut ExtCtxt, trait_span: Span, substr: &Substructure) -> @Expr {
let (lsb0, f)= match substr.nonself_args {
[l, f] => (l, f),
_ => cx.span_bug(trait_span, "Incorrect number of arguments in `deriving(IterBytes)`")
_ => cx.span_bug(trait_span, "incorrect number of arguments in `deriving(IterBytes)`")
};
// Build the "explicitly borrowed" stack closure, "|_buf| f(_buf)".
let blk_arg = cx.ident_of("_buf");
@ -82,7 +82,7 @@ fn iter_bytes_substructure(cx: &mut ExtCtxt, trait_span: Span, substr: &Substruc
fields = fs;
}
_ => cx.span_bug(trait_span, "Impossible substructure in `deriving(IterBytes)`")
_ => cx.span_bug(trait_span, "impossible substructure in `deriving(IterBytes)`")
}
for &FieldInfo { self_, span, .. } in fields.iter() {

View File

@ -68,7 +68,7 @@ pub fn expand_deriving_from_primitive(cx: &mut ExtCtxt,
fn cs_from(name: &str, cx: &mut ExtCtxt, trait_span: Span, substr: &Substructure) -> @Expr {
let n = match substr.nonself_args {
[n] => n,
_ => cx.span_bug(trait_span, "Incorrect number of arguments in `deriving(FromPrimitive)`")
_ => cx.span_bug(trait_span, "incorrect number of arguments in `deriving(FromPrimitive)`")
};
match *substr.fields {

View File

@ -178,8 +178,8 @@ impl<'a> Ty<'a> {
Literal(ref p) => {
p.to_path(cx, span, self_ty, self_generics)
}
Ptr(..) => { cx.span_bug(span, "Pointer in a path in generic `deriving`") }
Tuple(..) => { cx.span_bug(span, "Tuple in a path in generic `deriving`") }
Ptr(..) => { cx.span_bug(span, "pointer in a path in generic `deriving`") }
Tuple(..) => { cx.span_bug(span, "tuple in a path in generic `deriving`") }
}
}
}

View File

@ -257,7 +257,7 @@ pub mod rt {
match res {
Some(ast) => ast,
None => {
error!("Parse error");
error!("parse error");
fail!()
}
}
@ -589,7 +589,7 @@ fn expand_tts(cx: &ExtCtxt, sp: Span, tts: &[ast::TokenTree])
let cx_expr = p.parse_expr();
if !p.eat(&token::COMMA) {
p.fatal("Expected token `,`");
p.fatal("expected token `,`");
}
let tts = p.parse_all_token_trees();

View File

@ -49,7 +49,7 @@ pub fn find_macro_registrar(diagnostic: @diagnostic::SpanHandler,
})
},
_ => {
diagnostic.handler().err("Multiple macro registration functions found");
diagnostic.handler().err("multiple macro registration functions found");
for &(_, span) in ctx.registrars.iter() {
diagnostic.span_note(span, "one is here");
}

View File

@ -185,7 +185,7 @@ pub fn nameize(p_s: @ParseSess, ms: &[Matcher], res: &[@NamedMatch])
if ret_val.contains_key(bind_name) {
let string = token::get_ident(bind_name.name);
p_s.span_diagnostic
.span_fatal(sp, "Duplicated bind name: " + string.get())
.span_fatal(sp, "duplicated bind name: " + string.get())
}
ret_val.insert(*bind_name, res[idx]);
}
@ -441,6 +441,6 @@ pub fn parse_nt(p: &mut Parser, name: &str) -> Nonterminal {
res
}
"matchers" => token::NtMatchers(p.parse_matchers()),
_ => p.fatal(~"Unsupported builtin nonterminal parser: " + name)
_ => p.fatal(~"unsupported builtin nonterminal parser: " + name)
}
}

View File

@ -148,7 +148,7 @@ fn lis_merge(lhs: LockstepIterSize, rhs: LockstepIterSize) -> LockstepIterSize {
LisConstraint(r_len, ref r_id) => {
let l_n = token::get_ident(l_id.name);
let r_n = token::get_ident(r_id.name);
LisContradiction(format!("Inconsistent lockstep iteration: \
LisContradiction(format!("inconsistent lockstep iteration: \
'{}' has {} items, but '{}' has {}",
l_n.get(), l_len, r_n.get(), r_len))
}

View File

@ -85,7 +85,7 @@ impl<T> OptVec<T> {
pub fn get<'a>(&'a self, i: uint) -> &'a T {
match *self {
Empty => fail!("Invalid index {}", i),
Empty => fail!("invalid index {}", i),
Vec(ref v) => &v[i]
}
}
@ -103,7 +103,7 @@ impl<T> OptVec<T> {
pub fn swap_remove(&mut self, index: uint) {
match *self {
Empty => { fail!("Index out of bounds"); }
Empty => { fail!("index out of bounds"); }
Vec(ref mut v) => {
assert!(index < v.len());
v.swap_remove(index);

View File

@ -76,7 +76,7 @@ impl ParserObsoleteMethods for Parser {
let (kind_str, desc) = match kind {
ObsoleteSwap => (
"swap",
"Use std::util::{swap, replace} instead"
"use std::util::{swap, replace} instead"
),
ObsoleteUnsafeBlock => (
"non-standalone unsafe block",

View File

@ -431,7 +431,7 @@ impl Parser {
&& self.look_ahead(1, |t| *t == token::RBRACE) {
// matched; signal non-fatal error and recover.
self.span_err(self.span,
"Unit-like struct construction is written with no trailing `{ }`");
"unit-like struct construction is written with no trailing `{ }`");
self.eat(&token::LBRACE);
self.eat(&token::RBRACE);
true
@ -1601,7 +1601,7 @@ impl Parser {
};
}
_ => {
self.fatal(format!("Expected a lifetime name"));
self.fatal(format!("expected a lifetime name"));
}
}
}
@ -2137,7 +2137,7 @@ impl Parser {
}
// There shouldn't really be a span, but it's easier for the test runner
// if we give it one
self.fatal("This file contains an un-closed delimiter ");
self.fatal("this file contains an un-closed delimiter ");
}
token::LPAREN | token::LBRACE | token::LBRACKET => {
let close_delim = token::flip_delimiter(&self.token);
@ -3957,7 +3957,7 @@ impl Parser {
}
if fields.len() == 0 {
let string = get_ident_interner().get(class_name.name);
self.fatal(format!("Unit-like struct definition should be written as `struct {};`",
self.fatal(format!("unit-like struct definition should be written as `struct {};`",
string.as_slice()));
}
self.bump();

View File

@ -68,7 +68,7 @@ impl<T> SmallVector<T> {
match *self {
One(ref v) if idx == 0 => v,
Many(ref vs) => &vs[idx],
_ => fail!("Out of bounds access")
_ => fail!("out of bounds access")
}
}

View File

@ -19,4 +19,4 @@ fn bar() {
while (i < 1000) {}
}
fn main() {} //~ ERROR This file contains an un-closed delimiter
fn main() {} //~ ERROR this file contains an un-closed delimiter

View File

@ -20,16 +20,16 @@ fn main()
let _z = match g(1, 2) {
g(x, x) => { info!("{:?}", x + x); }
//~^ ERROR Identifier `x` is bound more than once in the same pattern
//~^ ERROR identifier `x` is bound more than once in the same pattern
};
let _z = match i(l(1, 2), m(3, 4)) {
i(l(x, _), m(_, x)) //~ ERROR Identifier `x` is bound more than once in the same pattern
i(l(x, _), m(_, x)) //~ ERROR identifier `x` is bound more than once in the same pattern
=> { error!("{:?}", x + x); }
};
let _z = match (1, 2) {
(x, x) => { x } //~ ERROR Identifier `x` is bound more than once in the same pattern
(x, x) => { x } //~ ERROR identifier `x` is bound more than once in the same pattern
};
}

View File

@ -10,7 +10,7 @@
use std::cmp::Eq;
trait Hahaha: Eq + Eq + Eq + Eq + Eq + //~ ERROR Duplicate supertrait
trait Hahaha: Eq + Eq + Eq + Eq + Eq + //~ ERROR duplicate supertrait
Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq +
Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq +
Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq +

View File

@ -8,7 +8,7 @@
// option. This file may not be copied, modified, or distributed
// except according to those terms.
// error-pattern: Multiple macro registration functions found
// error-pattern: multiple macro registration functions found
#[feature(macro_registrar)];

View File

@ -13,6 +13,6 @@
fn main() {
fn bar(n: int) {
let _x: [int, ..n];
//~^ ERROR expected constant expr for vector length: Non-constant path in constant expr
//~^ ERROR expected constant expr for vector length: non-constant path in constant expr
}
}

View File

@ -12,7 +12,7 @@ struct Foo;
fn f2() {
let _end_stmt = Foo { };
//~^ ERROR: Unit-like struct construction is written with no trailing `{ }`
//~^ ERROR: unit-like struct construction is written with no trailing `{ }`
}
fn main() {}

View File

@ -12,7 +12,7 @@ struct Foo;
fn g3() {
let _mid_tuple = (Foo { }, 2);
//~^ ERROR: Unit-like struct construction is written with no trailing `{ }`
//~^ ERROR: unit-like struct construction is written with no trailing `{ }`
}
fn main() {}

View File

@ -12,7 +12,7 @@ struct Foo;
fn h4() {
let _end_of_tuple = (3, Foo { });
//~^ ERROR: Unit-like struct construction is written with no trailing `{ }`
//~^ ERROR: unit-like struct construction is written with no trailing `{ }`
}
fn main() {}

View File

@ -12,7 +12,7 @@ struct Foo;
fn i5() {
let _end_of_block = { Foo { } };
//~^ ERROR: Unit-like struct construction is written with no trailing `{ }`
//~^ ERROR: unit-like struct construction is written with no trailing `{ }`
}
fn main() {}

View File

@ -9,6 +9,6 @@
// except according to those terms.
struct Foo {}
//~^ ERROR: Unit-like struct definition should be written as `struct Foo;`
//~^ ERROR: unit-like struct definition should be written as `struct Foo;`
fn main() {}

View File

@ -9,5 +9,5 @@
// except according to those terms.
fn main() {
let vec = bytes!('λ'); //~ ERROR Non-ascii char literal in bytes!
let vec = bytes!('λ'); //~ ERROR non-ascii char literal in bytes!
}

View File

@ -9,5 +9,5 @@
// except according to those terms.
fn main() {
let vec = bytes!(foo); //~ ERROR Non-literal in bytes!
let vec = bytes!(foo); //~ ERROR non-literal in bytes!
}

View File

@ -9,5 +9,5 @@
// except according to those terms.
fn main() {
let vec = bytes!(1024); //~ ERROR Too large integer literal in bytes!
let vec = bytes!(1024); //~ ERROR too large integer literal in bytes!
}

View File

@ -9,5 +9,5 @@
// except according to those terms.
fn main() {
let vec = bytes!(1024u8); //~ ERROR Too large u8 literal in bytes!
let vec = bytes!(1024u8); //~ ERROR too large u8 literal in bytes!
}

View File

@ -9,5 +9,5 @@
// except according to those terms.
fn main() {
let vec = bytes!(-1024); //~ ERROR Non-literal in bytes
let vec = bytes!(-1024); //~ ERROR non-literal in bytes
}

View File

@ -9,5 +9,5 @@
// except according to those terms.
fn main() {
let vec = bytes!(-1024u8); //~ ERROR Non-literal in bytes
let vec = bytes!(-1024u8); //~ ERROR non-literal in bytes
}

View File

@ -9,5 +9,5 @@
// except according to those terms.
fn main() {
let vec = bytes!(45f64); //~ ERROR Unsupported literal in bytes!
let vec = bytes!(45f64); //~ ERROR unsupported literal in bytes!
}