mirror of
https://github.com/rust-lang/rust.git
synced 2026-05-07 09:13:07 +03:00
Fixed error starting with uppercase
Error messages cleaned in librustc/middle Error messages cleaned in libsyntax Error messages cleaned in libsyntax more agressively Error messages cleaned in librustc more aggressively Fixed affected tests Fixed other failing tests Last failing tests fixed
This commit is contained in:
@@ -76,7 +76,7 @@ pub fn WriteOutputFile(
|
||||
let result = llvm::LLVMRustWriteOutputFile(
|
||||
Target, PM, M, Output, FileType);
|
||||
if !result {
|
||||
llvm_err(sess, ~"Could not write output");
|
||||
llvm_err(sess, ~"could not write output");
|
||||
}
|
||||
})
|
||||
}
|
||||
@@ -189,7 +189,7 @@ pub fn run_passes(sess: Session,
|
||||
for pass in sess.opts.custom_passes.iter() {
|
||||
pass.with_c_str(|s| {
|
||||
if !llvm::LLVMRustAddPass(mpm, s) {
|
||||
sess.warn(format!("Unknown pass {}, ignoring", *pass));
|
||||
sess.warn(format!("unknown pass {}, ignoring", *pass));
|
||||
}
|
||||
})
|
||||
}
|
||||
@@ -876,11 +876,11 @@ fn link_binary_output(sess: Session,
|
||||
let obj_is_writeable = is_writeable(&obj_filename);
|
||||
let out_is_writeable = is_writeable(&out_filename);
|
||||
if !out_is_writeable {
|
||||
sess.fatal(format!("Output file {} is not writeable -- check its permissions.",
|
||||
sess.fatal(format!("output file {} is not writeable -- check its permissions.",
|
||||
out_filename.display()));
|
||||
}
|
||||
else if !obj_is_writeable {
|
||||
sess.fatal(format!("Object file {} is not writeable -- check its permissions.",
|
||||
sess.fatal(format!("object file {} is not writeable -- check its permissions.",
|
||||
obj_filename.display()));
|
||||
}
|
||||
|
||||
|
||||
@@ -308,7 +308,7 @@ pub fn debugging_opt(&self, opt: u64) -> bool {
|
||||
// This exists to help with refactoring to eliminate impossible
|
||||
// cases later on
|
||||
pub fn impossible_case(&self, sp: Span, msg: &str) -> ! {
|
||||
self.span_bug(sp, format!("Impossible case reached: {}", msg));
|
||||
self.span_bug(sp, format!("impossible case reached: {}", msg));
|
||||
}
|
||||
pub fn verbose(&self) -> bool { self.debugging_opt(VERBOSE) }
|
||||
pub fn time_passes(&self) -> bool { self.debugging_opt(TIME_PASSES) }
|
||||
|
||||
@@ -527,7 +527,7 @@ fn parse_sig(st: &mut PState, conv: conv_did) -> ty::FnSig {
|
||||
let variadic = match next(st) {
|
||||
'V' => true,
|
||||
'N' => false,
|
||||
r => fail!(format!("Bad variadic: {}", r)),
|
||||
r => fail!(format!("bad variadic: {}", r)),
|
||||
};
|
||||
let ret_ty = parse_ty(st, |x,y| conv(x,y));
|
||||
ty::FnSig {binder_id: id,
|
||||
|
||||
@@ -195,7 +195,7 @@ fn enc_region(w: &mut MemWriter, cx: @ctxt, r: ty::Region) {
|
||||
}
|
||||
ty::ReInfer(_) => {
|
||||
// these should not crop up after typeck
|
||||
cx.diag.handler().bug("Cannot encode region variables");
|
||||
cx.diag.handler().bug("cannot encode region variables");
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -320,7 +320,7 @@ fn enc_sty(w: &mut MemWriter, cx: @ctxt, st: &ty::sty) {
|
||||
enc_bare_fn_ty(w, cx, f);
|
||||
}
|
||||
ty::ty_infer(_) => {
|
||||
cx.diag.handler().bug("Cannot encode inference variable types");
|
||||
cx.diag.handler().bug("cannot encode inference variable types");
|
||||
}
|
||||
ty::ty_param(param_ty {idx: id, def_id: did}) => {
|
||||
mywrite!(w, "p{}|{}", (cx.ds)(did), id);
|
||||
@@ -334,7 +334,7 @@ fn enc_sty(w: &mut MemWriter, cx: @ctxt, st: &ty::sty) {
|
||||
enc_substs(w, cx, substs);
|
||||
mywrite!(w, "]");
|
||||
}
|
||||
ty::ty_err => fail!("Shouldn't encode error type")
|
||||
ty::ty_err => fail!("shouldn't encode error type")
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@@ -215,7 +215,7 @@ fn check_root(&self,
|
||||
// the check above should fail for anything is not ReScope
|
||||
self.bccx.tcx.sess.span_bug(
|
||||
cmt_base.span,
|
||||
format!("Cannot issue root for scope region: {:?}",
|
||||
format!("cannot issue root for scope region: {:?}",
|
||||
self.loan_region));
|
||||
}
|
||||
};
|
||||
|
||||
@@ -496,7 +496,7 @@ pub fn guarantee_valid(&mut self,
|
||||
ty::ReInfer(..) => {
|
||||
self.tcx().sess.span_bug(
|
||||
cmt.span,
|
||||
format!("Invalid borrow lifetime: {:?}", loan_region));
|
||||
format!("invalid borrow lifetime: {:?}", loan_region));
|
||||
}
|
||||
};
|
||||
debug!("loan_scope = {:?}", loan_scope);
|
||||
@@ -820,7 +820,7 @@ pub fn vec_slice_info(&self, pat: &ast::Pat, slice_ty: ty::t)
|
||||
_ => {
|
||||
self.tcx().sess.span_bug(
|
||||
pat.span,
|
||||
format!("Type of slice pattern is not a slice"));
|
||||
format!("type of slice pattern is not a slice"));
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -769,14 +769,14 @@ pub fn append_loan_path_to_str(&self,
|
||||
}
|
||||
_ => {
|
||||
self.tcx.sess.bug(
|
||||
format!("Loan path LpVar({:?}) maps to {:?}, not local",
|
||||
format!("loan path LpVar({:?}) maps to {:?}, not local",
|
||||
id, pat));
|
||||
}
|
||||
}
|
||||
}
|
||||
r => {
|
||||
self.tcx.sess.bug(
|
||||
format!("Loan path LpVar({:?}) maps to {:?}, not local",
|
||||
format!("loan path LpVar({:?}) maps to {:?}, not local",
|
||||
id, r));
|
||||
}
|
||||
}
|
||||
|
||||
@@ -494,7 +494,7 @@ fn add_gen_kills(&self,
|
||||
dfcx_assign.add_kill(kill_id, assignment_index);
|
||||
}
|
||||
LpExtend(..) => {
|
||||
tcx.sess.bug("Var assignment for non var path");
|
||||
tcx.sess.bug("var assignment for non var path");
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -507,13 +507,13 @@ fn find_scope(&self,
|
||||
}
|
||||
self.tcx.sess.span_bug(
|
||||
expr.span,
|
||||
format!("No loop scope for id {:?}", loop_id));
|
||||
format!("no loop scope for id {:?}", loop_id));
|
||||
}
|
||||
|
||||
r => {
|
||||
self.tcx.sess.span_bug(
|
||||
expr.span,
|
||||
format!("Bad entry `{:?}` in def_map for label", r));
|
||||
format!("bad entry `{:?}` in def_map for label", r));
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -971,7 +971,7 @@ fn check_legality_of_move_bindings(cx: &MatchCheckCtxt,
|
||||
_ => {
|
||||
cx.tcx.sess.span_bug(
|
||||
p.span,
|
||||
format!("Binding pattern {} is \
|
||||
format!("binding pattern {} is \
|
||||
not an identifier: {:?}",
|
||||
p.id, p.node));
|
||||
}
|
||||
|
||||
@@ -342,8 +342,8 @@ fn fromb(b: bool) -> Result<const_val, ~str> { Ok(const_int(b as i64)) }
|
||||
Ok(const_float(f)) => Ok(const_float(-f)),
|
||||
Ok(const_int(i)) => Ok(const_int(-i)),
|
||||
Ok(const_uint(i)) => Ok(const_uint(-i)),
|
||||
Ok(const_str(_)) => Err(~"Negate on string"),
|
||||
Ok(const_bool(_)) => Err(~"Negate on boolean"),
|
||||
Ok(const_str(_)) => Err(~"negate on string"),
|
||||
Ok(const_bool(_)) => Err(~"negate on boolean"),
|
||||
ref err => ((*err).clone())
|
||||
}
|
||||
}
|
||||
@@ -352,7 +352,7 @@ fn fromb(b: bool) -> Result<const_val, ~str> { Ok(const_int(b as i64)) }
|
||||
Ok(const_int(i)) => Ok(const_int(!i)),
|
||||
Ok(const_uint(i)) => Ok(const_uint(!i)),
|
||||
Ok(const_bool(b)) => Ok(const_bool(!b)),
|
||||
_ => Err(~"Not on float or string")
|
||||
_ => Err(~"not on float or string")
|
||||
}
|
||||
}
|
||||
ExprBinary(_, op, a, b) => {
|
||||
@@ -371,7 +371,7 @@ fn fromb(b: bool) -> Result<const_val, ~str> { Ok(const_int(b as i64)) }
|
||||
BiNe => fromb(a != b),
|
||||
BiGe => fromb(a >= b),
|
||||
BiGt => fromb(a > b),
|
||||
_ => Err(~"Can't do this op on floats")
|
||||
_ => Err(~"can't do this op on floats")
|
||||
}
|
||||
}
|
||||
(Ok(const_int(a)), Ok(const_int(b))) => {
|
||||
@@ -423,14 +423,14 @@ fn fromb(b: bool) -> Result<const_val, ~str> { Ok(const_int(b as i64)) }
|
||||
match op {
|
||||
BiShl => Ok(const_int(a << b)),
|
||||
BiShr => Ok(const_int(a >> b)),
|
||||
_ => Err(~"Can't do this op on an int and uint")
|
||||
_ => Err(~"can't do this op on an int and uint")
|
||||
}
|
||||
}
|
||||
(Ok(const_uint(a)), Ok(const_int(b))) => {
|
||||
match op {
|
||||
BiShl => Ok(const_uint(a << b)),
|
||||
BiShr => Ok(const_uint(a >> b)),
|
||||
_ => Err(~"Can't do this op on a uint and int")
|
||||
_ => Err(~"can't do this op on a uint and int")
|
||||
}
|
||||
}
|
||||
(Ok(const_bool(a)), Ok(const_bool(b))) => {
|
||||
@@ -442,10 +442,10 @@ fn fromb(b: bool) -> Result<const_val, ~str> { Ok(const_int(b as i64)) }
|
||||
BiBitOr => a | b,
|
||||
BiEq => a == b,
|
||||
BiNe => a != b,
|
||||
_ => return Err(~"Can't do this op on bools")
|
||||
_ => return Err(~"can't do this op on bools")
|
||||
}))
|
||||
}
|
||||
_ => Err(~"Bad operands for binary")
|
||||
_ => Err(~"bad operands for binary")
|
||||
}
|
||||
}
|
||||
ExprCast(base, target_ty) => {
|
||||
@@ -456,7 +456,7 @@ fn fromb(b: bool) -> Result<const_val, ~str> { Ok(const_int(b as i64)) }
|
||||
.or_else(|| astconv::ast_ty_to_prim_ty(tcx.ty_ctxt(), target_ty))
|
||||
.unwrap_or_else(|| tcx.ty_ctxt().sess.span_fatal(
|
||||
target_ty.span,
|
||||
format!("Target type not found for const cast")
|
||||
format!("target type not found for const cast")
|
||||
));
|
||||
|
||||
let base = eval_const_expr_partial(tcx, base);
|
||||
@@ -469,7 +469,7 @@ fn fromb(b: bool) -> Result<const_val, ~str> { Ok(const_int(b as i64)) }
|
||||
const_uint(u) => Ok(const_float(u as f64)),
|
||||
const_int(i) => Ok(const_float(i as f64)),
|
||||
const_float(f) => Ok(const_float(f)),
|
||||
_ => Err(~"Can't cast float to str"),
|
||||
_ => Err(~"can't cast float to str"),
|
||||
}
|
||||
}
|
||||
ty::ty_uint(_) => {
|
||||
@@ -477,7 +477,7 @@ fn fromb(b: bool) -> Result<const_val, ~str> { Ok(const_int(b as i64)) }
|
||||
const_uint(u) => Ok(const_uint(u)),
|
||||
const_int(i) => Ok(const_uint(i as u64)),
|
||||
const_float(f) => Ok(const_uint(f as u64)),
|
||||
_ => Err(~"Can't cast str to uint"),
|
||||
_ => Err(~"can't cast str to uint"),
|
||||
}
|
||||
}
|
||||
ty::ty_int(_) | ty::ty_bool => {
|
||||
@@ -485,10 +485,10 @@ fn fromb(b: bool) -> Result<const_val, ~str> { Ok(const_int(b as i64)) }
|
||||
const_uint(u) => Ok(const_int(u as i64)),
|
||||
const_int(i) => Ok(const_int(i)),
|
||||
const_float(f) => Ok(const_int(f as i64)),
|
||||
_ => Err(~"Can't cast str to int"),
|
||||
_ => Err(~"can't cast str to int"),
|
||||
}
|
||||
}
|
||||
_ => Err(~"Can't cast this type")
|
||||
_ => Err(~"can't cast this type")
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -496,14 +496,14 @@ fn fromb(b: bool) -> Result<const_val, ~str> { Ok(const_int(b as i64)) }
|
||||
ExprPath(_) => {
|
||||
match lookup_const(tcx.ty_ctxt(), e) {
|
||||
Some(actual_e) => eval_const_expr_partial(&tcx.ty_ctxt(), actual_e),
|
||||
None => Err(~"Non-constant path in constant expr")
|
||||
None => Err(~"non-constant path in constant expr")
|
||||
}
|
||||
}
|
||||
ExprLit(lit) => Ok(lit_to_const(lit)),
|
||||
// If we have a vstore, just keep going; it has to be a string
|
||||
ExprVstore(e, _) => eval_const_expr_partial(tcx, e),
|
||||
ExprParen(e) => eval_const_expr_partial(tcx, e),
|
||||
_ => Err(~"Unsupported constant expr")
|
||||
_ => Err(~"unsupported constant expr")
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@@ -879,7 +879,7 @@ fn find_scope<'a>(&self,
|
||||
None => {
|
||||
self.tcx().sess.span_bug(
|
||||
expr.span,
|
||||
format!("No loop scope for id {:?}", loop_id));
|
||||
format!("no loop scope for id {:?}", loop_id));
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -887,7 +887,7 @@ fn find_scope<'a>(&self,
|
||||
r => {
|
||||
self.tcx().sess.span_bug(
|
||||
expr.span,
|
||||
format!("Bad entry `{:?}` in def_map for label", r));
|
||||
format!("bad entry `{:?}` in def_map for label", r));
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -328,7 +328,7 @@ pub fn variable(&self, node_id: NodeId, span: Span) -> Variable {
|
||||
Some(&var) => var,
|
||||
None => {
|
||||
self.tcx.sess.span_bug(
|
||||
span, format!("No variable registered for id {}", node_id));
|
||||
span, format!("no variable registered for id {}", node_id));
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -627,7 +627,7 @@ pub fn live_node(&self, node_id: NodeId, span: Span) -> LiveNode {
|
||||
// code have to agree about which AST nodes are worth
|
||||
// creating liveness nodes for.
|
||||
self.tcx.sess.span_bug(
|
||||
span, format!("No live node registered for node {}",
|
||||
span, format!("no live node registered for node {}",
|
||||
node_id));
|
||||
}
|
||||
}
|
||||
@@ -759,7 +759,7 @@ pub fn find_loop_scope(&self,
|
||||
let def_map = self.tcx.def_map.borrow();
|
||||
match def_map.get().find(&id) {
|
||||
Some(&DefLabel(loop_id)) => loop_id,
|
||||
_ => self.tcx.sess.span_bug(sp, "Label on break/loop \
|
||||
_ => self.tcx.sess.span_bug(sp, "label on break/loop \
|
||||
doesn't refer to a loop")
|
||||
}
|
||||
}
|
||||
@@ -1152,7 +1152,7 @@ pub fn propagate_through_expr(&self, expr: @Expr, succ: LiveNode)
|
||||
match break_ln.get().find(&sc) {
|
||||
Some(&b) => b,
|
||||
None => self.tcx.sess.span_bug(expr.span,
|
||||
"Break to unknown label")
|
||||
"break to unknown label")
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1167,7 +1167,7 @@ pub fn propagate_through_expr(&self, expr: @Expr, succ: LiveNode)
|
||||
match cont_ln.get().find(&sc) {
|
||||
Some(&b) => b,
|
||||
None => self.tcx.sess.span_bug(expr.span,
|
||||
"Loop to unknown label")
|
||||
"loop to unknown label")
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@@ -548,7 +548,7 @@ pub fn cat_def(&self,
|
||||
_ => {
|
||||
self.tcx.sess.span_bug(
|
||||
span,
|
||||
format!("Upvar of non-closure {:?} - {}",
|
||||
format!("upvar of non-closure {:?} - {}",
|
||||
fn_node_id, ty.repr(self.tcx)));
|
||||
}
|
||||
}
|
||||
@@ -652,7 +652,7 @@ pub fn cat_deref<N:ast_node>(&self,
|
||||
None => {
|
||||
self.tcx.sess.span_bug(
|
||||
node.span(),
|
||||
format!("Explicit deref of non-derefable type: {}",
|
||||
format!("explicit deref of non-derefable type: {}",
|
||||
ty_to_str(self.tcx, base_cmt.ty)));
|
||||
}
|
||||
};
|
||||
@@ -745,7 +745,7 @@ pub fn cat_index<N:ast_node>(&self,
|
||||
None => {
|
||||
self.tcx.sess.span_bug(
|
||||
elt.span(),
|
||||
format!("Explicit index of non-index type `{}`",
|
||||
format!("explicit index of non-index type `{}`",
|
||||
ty_to_str(self.tcx, base_cmt.ty)));
|
||||
}
|
||||
};
|
||||
|
||||
@@ -142,7 +142,7 @@ pub fn encl_scope(&self, id: ast::NodeId) -> ast::NodeId {
|
||||
let scope_map = self.scope_map.borrow();
|
||||
match scope_map.get().find(&id) {
|
||||
Some(&r) => r,
|
||||
None => { fail!("No enclosing scope for id {}", id); }
|
||||
None => { fail!("no enclosing scope for id {}", id); }
|
||||
}
|
||||
}
|
||||
|
||||
@@ -154,7 +154,7 @@ pub fn var_scope(&self, var_id: ast::NodeId) -> ast::NodeId {
|
||||
let var_map = self.var_map.borrow();
|
||||
match var_map.get().find(&var_id) {
|
||||
Some(&r) => r,
|
||||
None => { fail!("No enclosing scope for id {}", var_id); }
|
||||
None => { fail!("no enclosing scope for id {}", var_id); }
|
||||
}
|
||||
}
|
||||
|
||||
@@ -554,7 +554,7 @@ fn resolve_local(visitor: &mut RegionResolutionVisitor,
|
||||
None => {
|
||||
visitor.sess.span_bug(
|
||||
local.span,
|
||||
"Local without enclosing block");
|
||||
"local without enclosing block");
|
||||
}
|
||||
};
|
||||
|
||||
|
||||
@@ -1378,7 +1378,7 @@ fn build_reduced_graph_for_item(&mut self,
|
||||
}
|
||||
match method_map.get().find_mut(name) {
|
||||
Some(s) => { s.insert(def_id); },
|
||||
_ => fail!("Can't happen"),
|
||||
_ => fail!("can't happen"),
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1696,7 +1696,7 @@ trait method '{}'",
|
||||
}
|
||||
match method_map.get().find_mut(name) {
|
||||
Some(s) => { s.insert(def_id); },
|
||||
_ => fail!("Can't happen"),
|
||||
_ => fail!("can't happen"),
|
||||
}
|
||||
}
|
||||
|
||||
@@ -3992,7 +3992,7 @@ fn resolve_struct(&mut self,
|
||||
self.resolve_error(field.span,
|
||||
format!("field `{}` is already declared", ident_str));
|
||||
self.session.span_note(prev_field.span,
|
||||
"Previously declared here");
|
||||
"previously declared here");
|
||||
},
|
||||
None => {
|
||||
ident_map.insert(ident, field);
|
||||
@@ -4476,7 +4476,7 @@ struct in scope",
|
||||
// in the same disjunct, which is an
|
||||
// error
|
||||
self.resolve_error(pattern.span,
|
||||
format!("Identifier `{}` is bound more \
|
||||
format!("identifier `{}` is bound more \
|
||||
than once in the same pattern",
|
||||
path_to_str(path, self.session
|
||||
.intr())));
|
||||
|
||||
@@ -297,7 +297,7 @@ fn insert_lifetime(&self,
|
||||
def: ast::DefRegion) {
|
||||
if lifetime_ref.id == ast::DUMMY_NODE_ID {
|
||||
self.sess.span_bug(lifetime_ref.span,
|
||||
"Lifetime reference not renumbered, \
|
||||
"lifetime reference not renumbered, \
|
||||
probably a bug in syntax::fold");
|
||||
}
|
||||
|
||||
|
||||
@@ -484,7 +484,7 @@ fn assert_is_binding_or_wild(bcx: &Block, p: @ast::Pat) {
|
||||
if !pat_is_binding_or_wild(bcx.tcx().def_map, p) {
|
||||
bcx.sess().span_bug(
|
||||
p.span,
|
||||
format!("Expected an identifier pattern but found p: {}",
|
||||
format!("expected an identifier pattern but found p: {}",
|
||||
p.repr(bcx.tcx())));
|
||||
}
|
||||
}
|
||||
|
||||
@@ -434,7 +434,7 @@ fn generic_type_of(cx: &CrateContext, r: &Repr, name: Option<&str>, sizing: bool
|
||||
Type::array(&Type::i64(), align_units),
|
||||
a if a.population_count() == 1 => Type::array(&Type::vector(&Type::i32(), a / 4),
|
||||
align_units),
|
||||
_ => fail!("Unsupported enum alignment: {:?}", align)
|
||||
_ => fail!("unsupported enum alignment: {:?}", align)
|
||||
};
|
||||
assert_eq!(machine::llalign_of_min(cx, pad_ty) as u64, align);
|
||||
assert_eq!(align % discr_size, 0);
|
||||
|
||||
@@ -103,7 +103,7 @@ fn datum_callee<'a>(bcx: &'a Block<'a>, expr: &ast::Expr) -> Callee<'a> {
|
||||
_ => {
|
||||
bcx.tcx().sess.span_bug(
|
||||
expr.span,
|
||||
format!("Type of callee is neither bare-fn nor closure: {}",
|
||||
format!("type of callee is neither bare-fn nor closure: {}",
|
||||
bcx.ty_to_str(datum.ty)));
|
||||
}
|
||||
}
|
||||
@@ -151,7 +151,7 @@ fn trans_def<'a>(bcx: &'a Block<'a>, def: ast::Def, ref_expr: &ast::Expr)
|
||||
ast::DefSelfTy(..) | ast::DefMethod(..) => {
|
||||
bcx.tcx().sess.span_bug(
|
||||
ref_expr.span,
|
||||
format!("Cannot translate def {:?} \
|
||||
format!("cannot translate def {:?} \
|
||||
to a callable thing!", def));
|
||||
}
|
||||
}
|
||||
|
||||
@@ -211,7 +211,7 @@ fn top_loop_scope(&self) -> ast::NodeId {
|
||||
_ => {}
|
||||
}
|
||||
}
|
||||
self.ccx.tcx.sess.bug("No loop scope found");
|
||||
self.ccx.tcx.sess.bug("no loop scope found");
|
||||
}
|
||||
|
||||
fn normal_exit_block(&'a self,
|
||||
@@ -337,7 +337,7 @@ fn schedule_clean_in_ast_scope(&self,
|
||||
}
|
||||
|
||||
self.ccx.tcx.sess.bug(
|
||||
format!("No cleanup scope {} found",
|
||||
format!("no cleanup scope {} found",
|
||||
ast_map::node_id_to_str(self.ccx.tcx.items, cleanup_scope,
|
||||
token::get_ident_interner())));
|
||||
}
|
||||
@@ -548,7 +548,7 @@ fn trans_cleanups_to_exit_scope(&'a self,
|
||||
|
||||
LoopExit(id, _) => {
|
||||
self.ccx.tcx.sess.bug(format!(
|
||||
"Cannot exit from scope {:?}, \
|
||||
"cannot exit from scope {:?}, \
|
||||
not in scope", id));
|
||||
}
|
||||
}
|
||||
|
||||
@@ -472,7 +472,7 @@ pub fn def(&self, nid: ast::NodeId) -> ast::Def {
|
||||
Some(&v) => v,
|
||||
None => {
|
||||
self.tcx().sess.bug(format!(
|
||||
"No def associated with node id {:?}", nid));
|
||||
"no def associated with node id {:?}", nid));
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -852,7 +852,7 @@ pub fn node_id_type_params(bcx: &Block, id: ast::NodeId) -> ~[ty::t] {
|
||||
|
||||
if !params.iter().all(|t| !ty::type_needs_infer(*t)) {
|
||||
bcx.sess().bug(
|
||||
format!("Type parameters for node {} include inference types: {}",
|
||||
format!("type parameters for node {} include inference types: {}",
|
||||
id, params.map(|t| bcx.ty_to_str(*t)).connect(",")));
|
||||
}
|
||||
|
||||
|
||||
@@ -146,14 +146,14 @@ fn const_deref(cx: &CrateContext, v: ValueRef, t: ty::t, explicit: bool)
|
||||
const_deref_newtype(cx, v, t)
|
||||
}
|
||||
_ => {
|
||||
cx.sess.bug(format!("Unexpected dereferenceable type {}",
|
||||
cx.sess.bug(format!("unexpected dereferenceable type {}",
|
||||
ty_to_str(cx.tcx, t)))
|
||||
}
|
||||
};
|
||||
(dv, mt.ty)
|
||||
}
|
||||
None => {
|
||||
cx.sess.bug(format!("Can't dereference const of type {}",
|
||||
cx.sess.bug(format!("can't dereference const of type {}",
|
||||
ty_to_str(cx.tcx, t)))
|
||||
}
|
||||
}
|
||||
|
||||
@@ -298,7 +298,7 @@ pub fn create_local_var_metadata(bcx: &Block, local: &ast::Local) {
|
||||
Some(datum) => datum,
|
||||
None => {
|
||||
bcx.tcx().sess.span_bug(span,
|
||||
format!("No entry in lllocals table for {:?}",
|
||||
format!("no entry in lllocals table for {:?}",
|
||||
node_id));
|
||||
}
|
||||
}
|
||||
@@ -440,7 +440,7 @@ pub fn create_argument_metadata(bcx: &Block, arg: &ast::Arg) {
|
||||
Some(v) => v,
|
||||
None => {
|
||||
bcx.tcx().sess.span_bug(span,
|
||||
format!("No entry in llargs table for {:?}",
|
||||
format!("no entry in llargs table for {:?}",
|
||||
node_id));
|
||||
}
|
||||
}
|
||||
|
||||
@@ -81,17 +81,17 @@ pub fn llvm_calling_convention(ccx: &CrateContext,
|
||||
match abi {
|
||||
RustIntrinsic => {
|
||||
// Intrinsics are emitted by monomorphic fn
|
||||
ccx.sess.bug(format!("Asked to register intrinsic fn"));
|
||||
ccx.sess.bug(format!("asked to register intrinsic fn"));
|
||||
}
|
||||
|
||||
Rust => {
|
||||
// FIXME(#3678) Implement linking to foreign fns with Rust ABI
|
||||
ccx.sess.unimpl(
|
||||
format!("Foreign functions with Rust ABI"));
|
||||
format!("foreign functions with Rust ABI"));
|
||||
}
|
||||
|
||||
// It's the ABI's job to select this, not us.
|
||||
System => ccx.sess.bug("System abi should be selected elsewhere"),
|
||||
System => ccx.sess.bug("system abi should be selected elsewhere"),
|
||||
|
||||
Stdcall => lib::llvm::X86StdcallCallConv,
|
||||
Fastcall => lib::llvm::X86FastcallCallConv,
|
||||
@@ -365,7 +365,7 @@ pub fn trans_foreign_mod(ccx: @CrateContext,
|
||||
(abis, (*path).clone())
|
||||
}
|
||||
_ => {
|
||||
fail!("Unable to find foreign item in tcx.items \
|
||||
fail!("unable to find foreign item in tcx.items \
|
||||
table.")
|
||||
}
|
||||
};
|
||||
|
||||
@@ -227,7 +227,7 @@ fn count_zeros_intrinsic(bcx: &Block, name: &'static str) {
|
||||
"acq" => lib::llvm::Acquire,
|
||||
"rel" => lib::llvm::Release,
|
||||
"acqrel" => lib::llvm::AcquireRelease,
|
||||
_ => ccx.sess.fatal("Unknown ordering in atomic intrinsic")
|
||||
_ => ccx.sess.fatal("unknown ordering in atomic intrinsic")
|
||||
}
|
||||
};
|
||||
|
||||
@@ -268,7 +268,7 @@ fn count_zeros_intrinsic(bcx: &Block, name: &'static str) {
|
||||
"min" => lib::llvm::Min,
|
||||
"umax" => lib::llvm::UMax,
|
||||
"umin" => lib::llvm::UMin,
|
||||
_ => ccx.sess.fatal("Unknown atomic operation")
|
||||
_ => ccx.sess.fatal("unknown atomic operation")
|
||||
};
|
||||
|
||||
let old = AtomicRMW(bcx, atom_op, get_param(decl, first_real_arg),
|
||||
|
||||
@@ -98,7 +98,7 @@ pub fn monomorphic_fn(ccx: @CrateContext,
|
||||
session::expect(
|
||||
ccx.sess,
|
||||
ccx.tcx.items.find(fn_id.node),
|
||||
|| format!("While monomorphizing {:?}, couldn't find it in the \
|
||||
|| format!("while monomorphizing {:?}, couldn't find it in the \
|
||||
item map (may have attempted to monomorphize an item \
|
||||
defined in a different crate?)", fn_id))
|
||||
};
|
||||
@@ -286,7 +286,7 @@ pub fn monomorphic_fn(ccx: @CrateContext,
|
||||
d
|
||||
}
|
||||
_ => {
|
||||
ccx.tcx.sess.bug(format!("Can't monomorphize a {:?}",
|
||||
ccx.tcx.sess.bug(format!("can't monomorphize a {:?}",
|
||||
map_node))
|
||||
}
|
||||
}
|
||||
@@ -310,7 +310,7 @@ pub fn monomorphic_fn(ccx: @CrateContext,
|
||||
ast_map::NodeBlock(..) |
|
||||
ast_map::NodeCalleeScope(..) |
|
||||
ast_map::NodeLocal(..) => {
|
||||
ccx.tcx.sess.bug(format!("Can't monomorphize a {:?}", map_node))
|
||||
ccx.tcx.sess.bug(format!("can't monomorphize a {:?}", map_node))
|
||||
}
|
||||
};
|
||||
|
||||
|
||||
@@ -95,7 +95,7 @@ pub fn visit(&mut self, ty_name: &str, args: &[ValueRef]) {
|
||||
let tcx = self.bcx.tcx();
|
||||
let mth_idx = ty::method_idx(
|
||||
tcx.sess.ident_of(~"visit_" + ty_name),
|
||||
*self.visitor_methods).expect(format!("Couldn't find visit method \
|
||||
*self.visitor_methods).expect(format!("couldn't find visit method \
|
||||
for {}", ty_name));
|
||||
let mth_ty =
|
||||
ty::mk_bare_fn(tcx, self.visitor_methods[mth_idx].fty.clone());
|
||||
|
||||
@@ -404,7 +404,7 @@ pub fn write_content<'a>(
|
||||
}
|
||||
_ => {
|
||||
bcx.tcx().sess.span_bug(content_expr.span,
|
||||
"Unexpected evec content");
|
||||
"unexpected evec content");
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -465,7 +465,7 @@ pub fn write_content<'a>(
|
||||
}
|
||||
_ => {
|
||||
bcx.tcx().sess.span_bug(content_expr.span,
|
||||
"Unexpected vec content");
|
||||
"unexpected vec content");
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -498,7 +498,7 @@ pub fn elements_required(bcx: &Block, content_expr: &ast::Expr) -> uint {
|
||||
ast::LitStr(ref s, _) => s.get().len(),
|
||||
_ => {
|
||||
bcx.tcx().sess.span_bug(content_expr.span,
|
||||
"Unexpected evec content")
|
||||
"unexpected evec content")
|
||||
}
|
||||
}
|
||||
},
|
||||
@@ -507,7 +507,7 @@ pub fn elements_required(bcx: &Block, content_expr: &ast::Expr) -> uint {
|
||||
ty::eval_repeat_count(&bcx.tcx(), count_expr)
|
||||
}
|
||||
_ => bcx.tcx().sess.span_bug(content_expr.span,
|
||||
"Unexpected vec content")
|
||||
"unexpected vec content")
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@@ -2101,7 +2101,7 @@ fn tc_ty(cx: ctxt,
|
||||
ty_type => TC::None,
|
||||
|
||||
ty_err => {
|
||||
cx.sess.bug("Asked to compute contents of error type");
|
||||
cx.sess.bug("asked to compute contents of error type");
|
||||
}
|
||||
};
|
||||
|
||||
@@ -2926,7 +2926,7 @@ pub fn adjust_ty(cx: ctxt,
|
||||
None => {
|
||||
cx.sess.span_bug(
|
||||
span,
|
||||
format!("The {}th autoderef failed: \
|
||||
format!("the {}th autoderef failed: \
|
||||
{}",
|
||||
i,
|
||||
ty_to_str(cx, adjusted_ty)));
|
||||
@@ -3132,7 +3132,7 @@ pub fn resolve_expr(tcx: ctxt, expr: &ast::Expr) -> ast::Def {
|
||||
Some(&def) => def,
|
||||
None => {
|
||||
tcx.sess.span_bug(expr.span, format!(
|
||||
"No def-map entry for expr {:?}", expr.id));
|
||||
"no def-map entry for expr {:?}", expr.id));
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -3209,7 +3209,7 @@ pub fn expr_kind(tcx: ctxt,
|
||||
|
||||
def => {
|
||||
tcx.sess.span_bug(expr.span, format!(
|
||||
"Uncategorized def for expr {:?}: {:?}",
|
||||
"uncategorized def for expr {:?}: {:?}",
|
||||
expr.id, def));
|
||||
}
|
||||
}
|
||||
@@ -3335,7 +3335,7 @@ pub fn field_idx_strict(tcx: ty::ctxt, name: ast::Name, fields: &[field])
|
||||
for f in fields.iter() { if f.ident.name == name { return i; } i += 1u; }
|
||||
let string = token::get_ident(name);
|
||||
tcx.sess.bug(format!(
|
||||
"No field named `{}` found in the list of fields `{:?}`",
|
||||
"no field named `{}` found in the list of fields `{:?}`",
|
||||
string.get(),
|
||||
fields.map(|f| tcx.sess.str_of(f.ident))));
|
||||
}
|
||||
@@ -3687,7 +3687,7 @@ fn lookup_locally_or_in_crate_store<V:Clone>(
|
||||
}
|
||||
|
||||
if def_id.crate == ast::LOCAL_CRATE {
|
||||
fail!("No def'n found for {:?} in tcx.{}", def_id, descr);
|
||||
fail!("no def'n found for {:?} in tcx.{}", def_id, descr);
|
||||
}
|
||||
let v = load_external();
|
||||
map.insert(def_id, v.clone());
|
||||
|
||||
@@ -789,7 +789,7 @@ fn search_for_autoptrd_method(&self, self_ty: ty::t, autoderefs: uint)
|
||||
ty_err => None,
|
||||
|
||||
ty_unboxed_vec(_) | ty_type | ty_infer(TyVar(_)) => {
|
||||
self.bug(format!("Unexpected type: {}",
|
||||
self.bug(format!("unexpected type: {}",
|
||||
self.ty_to_str(self_ty)));
|
||||
}
|
||||
}
|
||||
|
||||
@@ -517,9 +517,9 @@ pub fn check_no_duplicate_fields(tcx: ty::ctxt,
|
||||
let orig_sp = field_names.find(&id).map(|x| *x);
|
||||
match orig_sp {
|
||||
Some(orig_sp) => {
|
||||
tcx.sess.span_err(sp, format!("Duplicate field name {} in record type declaration",
|
||||
tcx.sess.span_err(sp, format!("duplicate field name {} in record type declaration",
|
||||
tcx.sess.str_of(id)));
|
||||
tcx.sess.span_note(orig_sp, "First declaration of this field occurred here");
|
||||
tcx.sess.span_note(orig_sp, "first declaration of this field occurred here");
|
||||
break;
|
||||
}
|
||||
None => {
|
||||
@@ -1006,7 +1006,7 @@ pub fn local_ty(&self, span: Span, nid: ast::NodeId) -> ty::t {
|
||||
None => {
|
||||
self.tcx().sess.span_bug(
|
||||
span,
|
||||
format!("No type for local variable {:?}", nid));
|
||||
format!("no type for local variable {:?}", nid));
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -1619,7 +1619,7 @@ fn check_method_argument_types(
|
||||
_ => {
|
||||
fcx.tcx().sess.span_bug(
|
||||
sp,
|
||||
format!("Method without bare fn type"));
|
||||
format!("method without bare fn type"));
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -473,7 +473,7 @@ fn check_expr_fn_block(rcx: &mut Rcx,
|
||||
_ => {
|
||||
tcx.sess.span_bug(
|
||||
expr.span,
|
||||
"Expected expr_fn_block");
|
||||
"expected expr_fn_block");
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -1189,7 +1189,7 @@ fn apply_autoderefs(
|
||||
None => {
|
||||
tcx.sess.span_bug(
|
||||
expr.span,
|
||||
format!("Autoderef but type not derefable: {}",
|
||||
format!("autoderef but type not derefable: {}",
|
||||
ty_to_str(tcx, ct.ty)));
|
||||
}
|
||||
}
|
||||
|
||||
@@ -430,7 +430,7 @@ pub fn ensure_supertraits(ccx: &CrateCtxt,
|
||||
if ty_trait_refs.iter().any(|other_trait| other_trait.def_id == trait_ref.def_id) {
|
||||
// This means a trait inherited from the same supertrait more
|
||||
// than once.
|
||||
tcx.sess.span_err(sp, "Duplicate supertrait in trait declaration");
|
||||
tcx.sess.span_err(sp, "duplicate supertrait in trait declaration");
|
||||
break;
|
||||
} else {
|
||||
ty_trait_refs.push(trait_ref);
|
||||
@@ -891,7 +891,7 @@ pub fn ty_of_item(ccx: &CrateCtxt, it: &ast::Item)
|
||||
ast::ItemTrait(..) => {
|
||||
tcx.sess.span_bug(
|
||||
it.span,
|
||||
format!("Invoked ty_of_item on trait"));
|
||||
format!("invoked ty_of_item on trait"));
|
||||
}
|
||||
ast::ItemStruct(_, ref generics) => {
|
||||
let ty_generics = ty_generics(ccx, generics, 0);
|
||||
|
||||
@@ -202,7 +202,7 @@ pub fn unpack_actual_value(&self, a: ty::t, f: |&ty::sty| -> CoerceResult)
|
||||
Err(e) => {
|
||||
self.get_ref().infcx.tcx.sess.span_bug(
|
||||
self.get_ref().trace.origin.span(),
|
||||
format!("Failed to resolve even without \
|
||||
format!("failed to resolve even without \
|
||||
any force options: {:?}", e));
|
||||
}
|
||||
}
|
||||
|
||||
@@ -528,7 +528,7 @@ pub fn var_ids<T:Combine>(this: &T,
|
||||
r => {
|
||||
this.infcx().tcx.sess.span_bug(
|
||||
this.trace().origin.span(),
|
||||
format!("Found non-region-vid: {:?}", r));
|
||||
format!("found non-region-vid: {:?}", r));
|
||||
}
|
||||
}).collect()
|
||||
}
|
||||
|
||||
@@ -233,7 +233,7 @@ pub fn new_bound(&self, binder_id: ast::NodeId) -> Region {
|
||||
self.bound_count.set(sc + 1);
|
||||
|
||||
if sc >= self.bound_count.get() {
|
||||
self.tcx.sess.bug("Rollover in RegionInference new_bound()");
|
||||
self.tcx.sess.bug("rollover in RegionInference new_bound()");
|
||||
}
|
||||
|
||||
ReLateBound(binder_id, BrFresh(sc))
|
||||
@@ -278,7 +278,7 @@ pub fn make_subregion(&self,
|
||||
(_, ReLateBound(..)) => {
|
||||
self.tcx.sess.span_bug(
|
||||
origin.span(),
|
||||
format!("Cannot relate bound region: {} <= {}",
|
||||
format!("cannot relate bound region: {} <= {}",
|
||||
sub.repr(self.tcx),
|
||||
sup.repr(self.tcx)));
|
||||
}
|
||||
@@ -351,7 +351,7 @@ pub fn resolve_var(&self, rid: RegionVid) -> ty::Region {
|
||||
let var_origins = self.var_origins.borrow();
|
||||
self.tcx.sess.span_bug(
|
||||
var_origins.get()[rid.to_uint()].span(),
|
||||
format!("Attempt to resolve region variable before \
|
||||
format!("attempt to resolve region variable before \
|
||||
values have been computed!"))
|
||||
}
|
||||
Some(ref values) => values[rid.to_uint()]
|
||||
@@ -544,7 +544,7 @@ fn lub_concrete_regions(&self, a: Region, b: Region) -> Region {
|
||||
(ReEarlyBound(..), _) |
|
||||
(_, ReEarlyBound(..)) => {
|
||||
self.tcx.sess.bug(
|
||||
format!("Cannot relate bound region: LUB({}, {})",
|
||||
format!("cannot relate bound region: LUB({}, {})",
|
||||
a.repr(self.tcx),
|
||||
b.repr(self.tcx)));
|
||||
}
|
||||
@@ -646,7 +646,7 @@ fn glb_concrete_regions(&self,
|
||||
(ReEarlyBound(..), _) |
|
||||
(_, ReEarlyBound(..)) => {
|
||||
self.tcx.sess.bug(
|
||||
format!("Cannot relate bound region: GLB({}, {})",
|
||||
format!("cannot relate bound region: GLB({}, {})",
|
||||
a.repr(self.tcx),
|
||||
b.repr(self.tcx)));
|
||||
}
|
||||
|
||||
@@ -97,7 +97,7 @@ pub fn lookup_item(&self, names: &[~str]) -> ast::node_id {
|
||||
return match search_mod(self, &self.crate.node.module, 0, names) {
|
||||
Some(id) => id,
|
||||
None => {
|
||||
fail!("No item found: `%s`", names.connect("::"));
|
||||
fail!("no item found: `%s`", names.connect("::"));
|
||||
}
|
||||
};
|
||||
|
||||
@@ -230,7 +230,7 @@ pub fn resolve_regions(exp_count: uint) {
|
||||
for msg in self.err_messages.iter() {
|
||||
debug!("Error encountered: %s", *msg);
|
||||
}
|
||||
format!("Resolving regions encountered %u errors but expected %u!",
|
||||
format!("resolving regions encountered %u errors but expected %u!",
|
||||
self.err_messages.len(),
|
||||
exp_count);
|
||||
}
|
||||
@@ -240,7 +240,7 @@ pub fn resolve_regions(exp_count: uint) {
|
||||
pub fn check_lub(&self, t1: ty::t, t2: ty::t, t_lub: ty::t) {
|
||||
match self.lub().tys(t1, t2) {
|
||||
Err(e) => {
|
||||
fail!("Unexpected error computing LUB: %?", e)
|
||||
fail!("unexpected error computing LUB: %?", e)
|
||||
}
|
||||
Ok(t) => {
|
||||
self.assert_eq(t, t_lub);
|
||||
@@ -262,7 +262,7 @@ pub fn check_glb(&self, t1: ty::t, t2: ty::t, t_glb: ty::t) {
|
||||
self.ty_to_str(t_glb));
|
||||
match self.glb().tys(t1, t2) {
|
||||
Err(e) => {
|
||||
fail!("Unexpected error computing LUB: %?", e)
|
||||
fail!("unexpected error computing LUB: %?", e)
|
||||
}
|
||||
Ok(t) => {
|
||||
self.assert_eq(t, t_glb);
|
||||
@@ -281,7 +281,7 @@ pub fn check_no_lub(&self, t1: ty::t, t2: ty::t) {
|
||||
match self.lub().tys(t1, t2) {
|
||||
Err(_) => {}
|
||||
Ok(t) => {
|
||||
fail!("Unexpected success computing LUB: %?", self.ty_to_str(t))
|
||||
fail!("unexpected success computing LUB: %?", self.ty_to_str(t))
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -291,7 +291,7 @@ pub fn check_no_glb(&self, t1: ty::t, t2: ty::t) {
|
||||
match self.glb().tys(t1, t2) {
|
||||
Err(_) => {}
|
||||
Ok(t) => {
|
||||
fail!("Unexpected success computing GLB: %?", self.ty_to_str(t))
|
||||
fail!("unexpected success computing GLB: %?", self.ty_to_str(t))
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -718,7 +718,7 @@ fn add_constraints_from_ty(&mut self,
|
||||
ty::ty_infer(..) | ty::ty_err |
|
||||
ty::ty_type | ty::ty_unboxed_vec(..) => {
|
||||
self.tcx().sess.bug(
|
||||
format!("Unexpected type encountered in \
|
||||
format!("unexpected type encountered in \
|
||||
variance inference: {}",
|
||||
ty.repr(self.tcx())));
|
||||
}
|
||||
@@ -804,7 +804,7 @@ fn add_constraints_from_region(&mut self,
|
||||
ty::ReEmpty => {
|
||||
// We don't expect to see anything but 'static or bound
|
||||
// regions when visiting member types or method types.
|
||||
self.tcx().sess.bug(format!("Unexpected region encountered in \
|
||||
self.tcx().sess.bug(format!("unexpected region encountered in \
|
||||
variance inference: {}",
|
||||
region.repr(self.tcx())));
|
||||
}
|
||||
|
||||
@@ -64,12 +64,12 @@ fn add_bytes_to_bits<T: Int + CheckedAdd + ToBits>(bits: T, bytes: T) -> T {
|
||||
let (new_high_bits, new_low_bits) = bytes.to_bits();
|
||||
|
||||
if new_high_bits > Zero::zero() {
|
||||
fail!("Numeric overflow occured.")
|
||||
fail!("numeric overflow occured.")
|
||||
}
|
||||
|
||||
match bits.checked_add(&new_low_bits) {
|
||||
Some(x) => return x,
|
||||
None => fail!("Numeric overflow occured.")
|
||||
None => fail!("numeric overflow occured.")
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@@ -323,7 +323,7 @@ fn cannot_combine(n: Abi, m: Abi) {
|
||||
(m == a && n == b));
|
||||
}
|
||||
None => {
|
||||
fail!("Invalid match not detected");
|
||||
fail!("invalid match not detected");
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -335,7 +335,7 @@ fn can_combine(n: Abi, m: Abi) {
|
||||
set.add(m);
|
||||
match set.check_valid() {
|
||||
Some((_, _)) => {
|
||||
fail!("Valid match declared invalid");
|
||||
fail!("valid match declared invalid");
|
||||
}
|
||||
None => {}
|
||||
}
|
||||
|
||||
@@ -40,7 +40,7 @@ pub fn expand_syntax_ext(cx: &mut ExtCtxt, sp: Span, tts: &[ast::TokenTree]) ->
|
||||
// u8 literal, push to vector expression
|
||||
ast::LitUint(v, ast::TyU8) => {
|
||||
if v > 0xFF {
|
||||
cx.span_err(expr.span, "Too large u8 literal in bytes!")
|
||||
cx.span_err(expr.span, "too large u8 literal in bytes!")
|
||||
} else {
|
||||
bytes.push(cx.expr_u8(expr.span, v as u8));
|
||||
}
|
||||
@@ -49,9 +49,9 @@ pub fn expand_syntax_ext(cx: &mut ExtCtxt, sp: Span, tts: &[ast::TokenTree]) ->
|
||||
// integer literal, push to vector expression
|
||||
ast::LitIntUnsuffixed(v) => {
|
||||
if v > 0xFF {
|
||||
cx.span_err(expr.span, "Too large integer literal in bytes!")
|
||||
cx.span_err(expr.span, "too large integer literal in bytes!")
|
||||
} else if v < 0 {
|
||||
cx.span_err(expr.span, "Negative integer literal in bytes!")
|
||||
cx.span_err(expr.span, "negative integer literal in bytes!")
|
||||
} else {
|
||||
bytes.push(cx.expr_u8(expr.span, v as u8));
|
||||
}
|
||||
@@ -62,14 +62,14 @@ pub fn expand_syntax_ext(cx: &mut ExtCtxt, sp: Span, tts: &[ast::TokenTree]) ->
|
||||
if char::from_u32(v).unwrap().is_ascii() {
|
||||
bytes.push(cx.expr_u8(expr.span, v as u8));
|
||||
} else {
|
||||
cx.span_err(expr.span, "Non-ascii char literal in bytes!")
|
||||
cx.span_err(expr.span, "non-ascii char literal in bytes!")
|
||||
}
|
||||
}
|
||||
|
||||
_ => cx.span_err(expr.span, "Unsupported literal in bytes!")
|
||||
_ => cx.span_err(expr.span, "unsupported literal in bytes!")
|
||||
},
|
||||
|
||||
_ => cx.span_err(expr.span, "Non-literal in bytes!")
|
||||
_ => cx.span_err(expr.span, "non-literal in bytes!")
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@@ -92,10 +92,10 @@ fn cs_clone(
|
||||
all_fields = af;
|
||||
},
|
||||
EnumNonMatching(..) => cx.span_bug(trait_span,
|
||||
format!("Non-matching enum variants in `deriving({})`",
|
||||
format!("non-matching enum variants in `deriving({})`",
|
||||
name)),
|
||||
StaticEnum(..) | StaticStruct(..) => cx.span_bug(trait_span,
|
||||
format!("Static method in `deriving({})`",
|
||||
format!("static method in `deriving({})`",
|
||||
name))
|
||||
}
|
||||
|
||||
|
||||
@@ -75,7 +75,7 @@ fn cs_op(less: bool, equal: bool, cx: &mut ExtCtxt, span: Span, substr: &Substru
|
||||
*/
|
||||
let other_f = match other_fs {
|
||||
[o_f] => o_f,
|
||||
_ => cx.span_bug(span, "Not exactly 2 arguments in `deriving(Ord)`")
|
||||
_ => cx.span_bug(span, "not exactly 2 arguments in `deriving(Ord)`")
|
||||
};
|
||||
|
||||
let cmp = cx.expr_binary(span, op, self_f, other_f);
|
||||
@@ -99,7 +99,7 @@ fn cs_op(less: bool, equal: bool, cx: &mut ExtCtxt, span: Span, substr: &Substru
|
||||
} else {
|
||||
self_var > other_var
|
||||
}),
|
||||
_ => cx.span_bug(span, "Not exactly 2 arguments in `deriving(Ord)`")
|
||||
_ => cx.span_bug(span, "not exactly 2 arguments in `deriving(Ord)`")
|
||||
}
|
||||
},
|
||||
cx, span, substr)
|
||||
|
||||
@@ -110,7 +110,7 @@ pub fn cs_cmp(cx: &mut ExtCtxt, span: Span,
|
||||
let order = ordering_const(cx, span, self_var.cmp(&other_var));
|
||||
cx.expr_path(order)
|
||||
}
|
||||
_ => cx.span_bug(span, "Not exactly 2 arguments in `deriving(TotalOrd)`")
|
||||
_ => cx.span_bug(span, "not exactly 2 arguments in `deriving(TotalOrd)`")
|
||||
}
|
||||
},
|
||||
cx, span, substr)
|
||||
|
||||
@@ -656,7 +656,7 @@ fn expand_struct_method_body(&self,
|
||||
}).collect()
|
||||
}
|
||||
[] => { trait_.cx.span_bug(trait_.span,
|
||||
"No self arguments to non-static method \
|
||||
"no self arguments to non-static method \
|
||||
in generic `deriving`") }
|
||||
};
|
||||
|
||||
@@ -840,7 +840,7 @@ fn build_enum_match(&self,
|
||||
let index = match matching {
|
||||
Some(i) => i,
|
||||
None => cx.span_bug(trait_.span,
|
||||
"Non-matching variants when required to \
|
||||
"non-matching variants when required to \
|
||||
be matching in generic `deriving`")
|
||||
};
|
||||
|
||||
@@ -965,7 +965,7 @@ fn summarise_struct(&self, struct_def: &StructDef) -> StaticFields {
|
||||
|
||||
match (just_spans.is_empty(), named_idents.is_empty()) {
|
||||
(false, false) => self.cx.span_bug(self.span,
|
||||
"A struct with named and unnamed \
|
||||
"a struct with named and unnamed \
|
||||
fields in generic `deriving`"),
|
||||
// named fields
|
||||
(_, false) => Named(named_idents),
|
||||
@@ -1019,7 +1019,7 @@ fn create_struct_pattern(&self,
|
||||
None
|
||||
}
|
||||
_ => {
|
||||
cx.span_bug(sp, "A struct with named and unnamed fields in `deriving`");
|
||||
cx.span_bug(sp, "a struct with named and unnamed fields in `deriving`");
|
||||
}
|
||||
};
|
||||
let path = cx.path_ident(sp, cx.ident_of(format!("{}_{}", prefix, i)));
|
||||
@@ -1116,7 +1116,7 @@ pub fn cs_fold(use_foldl: bool,
|
||||
*all_enums,
|
||||
substructure.nonself_args),
|
||||
StaticEnum(..) | StaticStruct(..) => {
|
||||
cx.span_bug(trait_span, "Static function in `deriving`")
|
||||
cx.span_bug(trait_span, "static function in `deriving`")
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -1154,7 +1154,7 @@ pub fn cs_same_method(f: |&mut ExtCtxt, Span, ~[@Expr]| -> @Expr,
|
||||
*all_enums,
|
||||
substructure.nonself_args),
|
||||
StaticEnum(..) | StaticStruct(..) => {
|
||||
cx.span_bug(trait_span, "Static function in `deriving`")
|
||||
cx.span_bug(trait_span, "static function in `deriving`")
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -48,7 +48,7 @@ pub fn expand_deriving_iter_bytes(cx: &mut ExtCtxt,
|
||||
fn iter_bytes_substructure(cx: &mut ExtCtxt, trait_span: Span, substr: &Substructure) -> @Expr {
|
||||
let (lsb0, f)= match substr.nonself_args {
|
||||
[l, f] => (l, f),
|
||||
_ => cx.span_bug(trait_span, "Incorrect number of arguments in `deriving(IterBytes)`")
|
||||
_ => cx.span_bug(trait_span, "incorrect number of arguments in `deriving(IterBytes)`")
|
||||
};
|
||||
// Build the "explicitly borrowed" stack closure, "|_buf| f(_buf)".
|
||||
let blk_arg = cx.ident_of("_buf");
|
||||
@@ -82,7 +82,7 @@ fn iter_bytes_substructure(cx: &mut ExtCtxt, trait_span: Span, substr: &Substruc
|
||||
|
||||
fields = fs;
|
||||
}
|
||||
_ => cx.span_bug(trait_span, "Impossible substructure in `deriving(IterBytes)`")
|
||||
_ => cx.span_bug(trait_span, "impossible substructure in `deriving(IterBytes)`")
|
||||
}
|
||||
|
||||
for &FieldInfo { self_, span, .. } in fields.iter() {
|
||||
|
||||
@@ -68,7 +68,7 @@ pub fn expand_deriving_from_primitive(cx: &mut ExtCtxt,
|
||||
fn cs_from(name: &str, cx: &mut ExtCtxt, trait_span: Span, substr: &Substructure) -> @Expr {
|
||||
let n = match substr.nonself_args {
|
||||
[n] => n,
|
||||
_ => cx.span_bug(trait_span, "Incorrect number of arguments in `deriving(FromPrimitive)`")
|
||||
_ => cx.span_bug(trait_span, "incorrect number of arguments in `deriving(FromPrimitive)`")
|
||||
};
|
||||
|
||||
match *substr.fields {
|
||||
|
||||
@@ -178,8 +178,8 @@ pub fn to_path(&self,
|
||||
Literal(ref p) => {
|
||||
p.to_path(cx, span, self_ty, self_generics)
|
||||
}
|
||||
Ptr(..) => { cx.span_bug(span, "Pointer in a path in generic `deriving`") }
|
||||
Tuple(..) => { cx.span_bug(span, "Tuple in a path in generic `deriving`") }
|
||||
Ptr(..) => { cx.span_bug(span, "pointer in a path in generic `deriving`") }
|
||||
Tuple(..) => { cx.span_bug(span, "tuple in a path in generic `deriving`") }
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -257,7 +257,7 @@ fn parse_item(&self, s: ~str) -> @ast::Item {
|
||||
match res {
|
||||
Some(ast) => ast,
|
||||
None => {
|
||||
error!("Parse error");
|
||||
error!("parse error");
|
||||
fail!()
|
||||
}
|
||||
}
|
||||
@@ -589,7 +589,7 @@ fn expand_tts(cx: &ExtCtxt, sp: Span, tts: &[ast::TokenTree])
|
||||
|
||||
let cx_expr = p.parse_expr();
|
||||
if !p.eat(&token::COMMA) {
|
||||
p.fatal("Expected token `,`");
|
||||
p.fatal("expected token `,`");
|
||||
}
|
||||
|
||||
let tts = p.parse_all_token_trees();
|
||||
|
||||
@@ -49,7 +49,7 @@ pub fn find_macro_registrar(diagnostic: @diagnostic::SpanHandler,
|
||||
})
|
||||
},
|
||||
_ => {
|
||||
diagnostic.handler().err("Multiple macro registration functions found");
|
||||
diagnostic.handler().err("multiple macro registration functions found");
|
||||
for &(_, span) in ctx.registrars.iter() {
|
||||
diagnostic.span_note(span, "one is here");
|
||||
}
|
||||
|
||||
@@ -185,7 +185,7 @@ fn n_rec(p_s: @ParseSess, m: &Matcher, res: &[@NamedMatch],
|
||||
if ret_val.contains_key(bind_name) {
|
||||
let string = token::get_ident(bind_name.name);
|
||||
p_s.span_diagnostic
|
||||
.span_fatal(sp, "Duplicated bind name: " + string.get())
|
||||
.span_fatal(sp, "duplicated bind name: " + string.get())
|
||||
}
|
||||
ret_val.insert(*bind_name, res[idx]);
|
||||
}
|
||||
@@ -441,6 +441,6 @@ pub fn parse_nt(p: &mut Parser, name: &str) -> Nonterminal {
|
||||
res
|
||||
}
|
||||
"matchers" => token::NtMatchers(p.parse_matchers()),
|
||||
_ => p.fatal(~"Unsupported builtin nonterminal parser: " + name)
|
||||
_ => p.fatal(~"unsupported builtin nonterminal parser: " + name)
|
||||
}
|
||||
}
|
||||
|
||||
@@ -148,7 +148,7 @@ fn lis_merge(lhs: LockstepIterSize, rhs: LockstepIterSize) -> LockstepIterSize {
|
||||
LisConstraint(r_len, ref r_id) => {
|
||||
let l_n = token::get_ident(l_id.name);
|
||||
let r_n = token::get_ident(r_id.name);
|
||||
LisContradiction(format!("Inconsistent lockstep iteration: \
|
||||
LisContradiction(format!("inconsistent lockstep iteration: \
|
||||
'{}' has {} items, but '{}' has {}",
|
||||
l_n.get(), l_len, r_n.get(), r_len))
|
||||
}
|
||||
|
||||
@@ -85,7 +85,7 @@ pub fn map_move<U>(self, op: |T| -> U) -> OptVec<U> {
|
||||
|
||||
pub fn get<'a>(&'a self, i: uint) -> &'a T {
|
||||
match *self {
|
||||
Empty => fail!("Invalid index {}", i),
|
||||
Empty => fail!("invalid index {}", i),
|
||||
Vec(ref v) => &v[i]
|
||||
}
|
||||
}
|
||||
@@ -103,7 +103,7 @@ pub fn len(&self) -> uint {
|
||||
|
||||
pub fn swap_remove(&mut self, index: uint) {
|
||||
match *self {
|
||||
Empty => { fail!("Index out of bounds"); }
|
||||
Empty => { fail!("index out of bounds"); }
|
||||
Vec(ref mut v) => {
|
||||
assert!(index < v.len());
|
||||
v.swap_remove(index);
|
||||
|
||||
@@ -76,7 +76,7 @@ fn obsolete(&mut self, sp: Span, kind: ObsoleteSyntax) {
|
||||
let (kind_str, desc) = match kind {
|
||||
ObsoleteSwap => (
|
||||
"swap",
|
||||
"Use std::util::{swap, replace} instead"
|
||||
"use std::util::{swap, replace} instead"
|
||||
),
|
||||
ObsoleteUnsafeBlock => (
|
||||
"non-standalone unsafe block",
|
||||
|
||||
@@ -431,7 +431,7 @@ pub fn check_for_erroneous_unit_struct_expecting(&mut self, expected: &[token::T
|
||||
&& self.look_ahead(1, |t| *t == token::RBRACE) {
|
||||
// matched; signal non-fatal error and recover.
|
||||
self.span_err(self.span,
|
||||
"Unit-like struct construction is written with no trailing `{ }`");
|
||||
"unit-like struct construction is written with no trailing `{ }`");
|
||||
self.eat(&token::LBRACE);
|
||||
self.eat(&token::RBRACE);
|
||||
true
|
||||
@@ -1601,7 +1601,7 @@ pub fn parse_lifetime(&mut self) -> ast::Lifetime {
|
||||
};
|
||||
}
|
||||
_ => {
|
||||
self.fatal(format!("Expected a lifetime name"));
|
||||
self.fatal(format!("expected a lifetime name"));
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -2137,7 +2137,7 @@ fn parse_any_tt_tok(p: &mut Parser) -> TokenTree {
|
||||
}
|
||||
// There shouldn't really be a span, but it's easier for the test runner
|
||||
// if we give it one
|
||||
self.fatal("This file contains an un-closed delimiter ");
|
||||
self.fatal("this file contains an un-closed delimiter ");
|
||||
}
|
||||
token::LPAREN | token::LBRACE | token::LBRACKET => {
|
||||
let close_delim = token::flip_delimiter(&self.token);
|
||||
@@ -3957,7 +3957,7 @@ fn parse_item_struct(&mut self) -> ItemInfo {
|
||||
}
|
||||
if fields.len() == 0 {
|
||||
let string = get_ident_interner().get(class_name.name);
|
||||
self.fatal(format!("Unit-like struct definition should be written as `struct {};`",
|
||||
self.fatal(format!("unit-like struct definition should be written as `struct {};`",
|
||||
string.as_slice()));
|
||||
}
|
||||
self.bump();
|
||||
|
||||
@@ -68,7 +68,7 @@ pub fn get<'a>(&'a self, idx: uint) -> &'a T {
|
||||
match *self {
|
||||
One(ref v) if idx == 0 => v,
|
||||
Many(ref vs) => &vs[idx],
|
||||
_ => fail!("Out of bounds access")
|
||||
_ => fail!("out of bounds access")
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@@ -19,4 +19,4 @@ fn bar() {
|
||||
while (i < 1000) {}
|
||||
}
|
||||
|
||||
fn main() {} //~ ERROR This file contains an un-closed delimiter
|
||||
fn main() {} //~ ERROR this file contains an un-closed delimiter
|
||||
|
||||
@@ -20,16 +20,16 @@ fn main()
|
||||
|
||||
let _z = match g(1, 2) {
|
||||
g(x, x) => { info!("{:?}", x + x); }
|
||||
//~^ ERROR Identifier `x` is bound more than once in the same pattern
|
||||
//~^ ERROR identifier `x` is bound more than once in the same pattern
|
||||
};
|
||||
|
||||
let _z = match i(l(1, 2), m(3, 4)) {
|
||||
i(l(x, _), m(_, x)) //~ ERROR Identifier `x` is bound more than once in the same pattern
|
||||
i(l(x, _), m(_, x)) //~ ERROR identifier `x` is bound more than once in the same pattern
|
||||
=> { error!("{:?}", x + x); }
|
||||
};
|
||||
|
||||
let _z = match (1, 2) {
|
||||
(x, x) => { x } //~ ERROR Identifier `x` is bound more than once in the same pattern
|
||||
(x, x) => { x } //~ ERROR identifier `x` is bound more than once in the same pattern
|
||||
};
|
||||
|
||||
}
|
||||
|
||||
@@ -10,7 +10,7 @@
|
||||
|
||||
use std::cmp::Eq;
|
||||
|
||||
trait Hahaha: Eq + Eq + Eq + Eq + Eq + //~ ERROR Duplicate supertrait
|
||||
trait Hahaha: Eq + Eq + Eq + Eq + Eq + //~ ERROR duplicate supertrait
|
||||
Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq +
|
||||
Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq +
|
||||
Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq + Eq +
|
||||
|
||||
@@ -8,7 +8,7 @@
|
||||
// option. This file may not be copied, modified, or distributed
|
||||
// except according to those terms.
|
||||
|
||||
// error-pattern: Multiple macro registration functions found
|
||||
// error-pattern: multiple macro registration functions found
|
||||
|
||||
#[feature(macro_registrar)];
|
||||
|
||||
|
||||
@@ -13,6 +13,6 @@
|
||||
fn main() {
|
||||
fn bar(n: int) {
|
||||
let _x: [int, ..n];
|
||||
//~^ ERROR expected constant expr for vector length: Non-constant path in constant expr
|
||||
//~^ ERROR expected constant expr for vector length: non-constant path in constant expr
|
||||
}
|
||||
}
|
||||
|
||||
@@ -12,7 +12,7 @@
|
||||
|
||||
fn f2() {
|
||||
let _end_stmt = Foo { };
|
||||
//~^ ERROR: Unit-like struct construction is written with no trailing `{ }`
|
||||
//~^ ERROR: unit-like struct construction is written with no trailing `{ }`
|
||||
}
|
||||
|
||||
fn main() {}
|
||||
|
||||
@@ -12,7 +12,7 @@
|
||||
|
||||
fn g3() {
|
||||
let _mid_tuple = (Foo { }, 2);
|
||||
//~^ ERROR: Unit-like struct construction is written with no trailing `{ }`
|
||||
//~^ ERROR: unit-like struct construction is written with no trailing `{ }`
|
||||
}
|
||||
|
||||
fn main() {}
|
||||
|
||||
@@ -12,7 +12,7 @@
|
||||
|
||||
fn h4() {
|
||||
let _end_of_tuple = (3, Foo { });
|
||||
//~^ ERROR: Unit-like struct construction is written with no trailing `{ }`
|
||||
//~^ ERROR: unit-like struct construction is written with no trailing `{ }`
|
||||
}
|
||||
|
||||
fn main() {}
|
||||
|
||||
@@ -12,7 +12,7 @@
|
||||
|
||||
fn i5() {
|
||||
let _end_of_block = { Foo { } };
|
||||
//~^ ERROR: Unit-like struct construction is written with no trailing `{ }`
|
||||
//~^ ERROR: unit-like struct construction is written with no trailing `{ }`
|
||||
}
|
||||
|
||||
fn main() {}
|
||||
|
||||
@@ -9,6 +9,6 @@
|
||||
// except according to those terms.
|
||||
|
||||
struct Foo {}
|
||||
//~^ ERROR: Unit-like struct definition should be written as `struct Foo;`
|
||||
//~^ ERROR: unit-like struct definition should be written as `struct Foo;`
|
||||
|
||||
fn main() {}
|
||||
|
||||
@@ -9,5 +9,5 @@
|
||||
// except according to those terms.
|
||||
|
||||
fn main() {
|
||||
let vec = bytes!('λ'); //~ ERROR Non-ascii char literal in bytes!
|
||||
let vec = bytes!('λ'); //~ ERROR non-ascii char literal in bytes!
|
||||
}
|
||||
|
||||
@@ -9,5 +9,5 @@
|
||||
// except according to those terms.
|
||||
|
||||
fn main() {
|
||||
let vec = bytes!(foo); //~ ERROR Non-literal in bytes!
|
||||
let vec = bytes!(foo); //~ ERROR non-literal in bytes!
|
||||
}
|
||||
|
||||
@@ -9,5 +9,5 @@
|
||||
// except according to those terms.
|
||||
|
||||
fn main() {
|
||||
let vec = bytes!(1024); //~ ERROR Too large integer literal in bytes!
|
||||
let vec = bytes!(1024); //~ ERROR too large integer literal in bytes!
|
||||
}
|
||||
|
||||
@@ -9,5 +9,5 @@
|
||||
// except according to those terms.
|
||||
|
||||
fn main() {
|
||||
let vec = bytes!(1024u8); //~ ERROR Too large u8 literal in bytes!
|
||||
let vec = bytes!(1024u8); //~ ERROR too large u8 literal in bytes!
|
||||
}
|
||||
|
||||
@@ -9,5 +9,5 @@
|
||||
// except according to those terms.
|
||||
|
||||
fn main() {
|
||||
let vec = bytes!(-1024); //~ ERROR Non-literal in bytes
|
||||
let vec = bytes!(-1024); //~ ERROR non-literal in bytes
|
||||
}
|
||||
|
||||
@@ -9,5 +9,5 @@
|
||||
// except according to those terms.
|
||||
|
||||
fn main() {
|
||||
let vec = bytes!(-1024u8); //~ ERROR Non-literal in bytes
|
||||
let vec = bytes!(-1024u8); //~ ERROR non-literal in bytes
|
||||
}
|
||||
|
||||
@@ -9,5 +9,5 @@
|
||||
// except according to those terms.
|
||||
|
||||
fn main() {
|
||||
let vec = bytes!(45f64); //~ ERROR Unsupported literal in bytes!
|
||||
let vec = bytes!(45f64); //~ ERROR unsupported literal in bytes!
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user