pub fn operator_prec(op: ast::BinOp_) -> usize {
match op {
// 'as' sits here with 12
- BiMul | BiDiv | BiRem => 11us,
- BiAdd | BiSub => 10us,
- BiShl | BiShr => 9us,
- BiBitAnd => 8us,
- BiBitXor => 7us,
- BiBitOr => 6us,
- BiLt | BiLe | BiGe | BiGt | BiEq | BiNe => 3us,
- BiAnd => 2us,
- BiOr => 1us
+ BiMul | BiDiv | BiRem => 11,
+ BiAdd | BiSub => 10,
+ BiShl | BiShr => 9,
+ BiBitAnd => 8,
+ BiBitXor => 7,
+ BiBitOr => 6,
+ BiLt | BiLe | BiGe | BiGt | BiEq | BiNe => 3,
+ BiAnd => 2,
+ BiOr => 1
}
}
/// Precedence of the `as` operator, which is a binary operator
/// not appearing in the prior table.
-pub const AS_PREC: usize = 12us;
+pub const AS_PREC: usize = 12;
pub fn empty_generics() -> Generics {
Generics {
let lo = self.lookup_char_pos(sp.lo);
let hi = self.lookup_char_pos(sp.hi);
let mut lines = Vec::new();
- for i in lo.line - 1us..hi.line as usize {
+ for i in lo.line - 1..hi.line as usize {
lines.push(i);
};
FileLines {file: lo.file, lines: lines}
let files = self.files.borrow();
let files = &*files;
let len = files.len();
- let mut a = 0us;
+ let mut a = 0;
let mut b = len;
- while b - a > 1us {
- let m = (a + b) / 2us;
+ while b - a > 1 {
+ let m = (a + b) / 2;
if files[m].start_pos > pos {
b = m;
} else {
let files = self.files.borrow();
let f = (*files)[idx].clone();
- let mut a = 0us;
+ let mut a = 0;
{
let lines = f.lines.borrow();
let mut b = lines.len();
- while b - a > 1us {
- let m = (a + b) / 2us;
+ while b - a > 1 {
+ let m = (a + b) / 2;
if (*lines)[m] > pos { b = m; } else { a = m; }
}
}
fn lookup_pos(&self, pos: BytePos) -> Loc {
let FileMapAndLine {fm: f, line: a} = self.lookup_line(pos);
- let line = a + 1us; // Line numbers start at 1
+ let line = a + 1; // Line numbers start at 1
let chpos = self.bytepos_to_file_charpos(pos);
let linebpos = (*f.lines.borrow())[a];
let linechpos = self.bytepos_to_file_charpos(linebpos);
assert_eq!(file_lines.file.name, "blork.rs");
assert_eq!(file_lines.lines.len(), 1);
- assert_eq!(file_lines.lines[0], 1us);
+ assert_eq!(file_lines.lines[0], 1);
}
#[test]
use term;
/// maximum number of lines we will print for each error; arbitrary.
-static MAX_LINES: usize = 6us;
+static MAX_LINES: usize = 6;
#[derive(Clone, Copy)]
pub enum RenderSpan {
self.bump_err_count();
}
pub fn bump_err_count(&self) {
- self.err_count.set(self.err_count.get() + 1us);
+ self.err_count.set(self.err_count.get() + 1);
}
pub fn err_count(&self) -> usize {
self.err_count.get()
}
pub fn has_errors(&self) -> bool {
- self.err_count.get() > 0us
+ self.err_count.get() > 0
}
pub fn abort_if_errors(&self) {
let s;
match self.err_count.get() {
- 0us => return,
- 1us => s = "aborting due to previous error".to_string(),
+ 0 => return,
+ 1 => s = "aborting due to previous error".to_string(),
_ => {
s = format!("aborting due to {} previous errors",
self.err_count.get());
let mut elided = false;
let mut display_lines = &lines.lines[];
if display_lines.len() > MAX_LINES {
- display_lines = &display_lines[0us..MAX_LINES];
+ display_lines = &display_lines[0..MAX_LINES];
elided = true;
}
// Print the offending lines
}
}
if elided {
- let last_line = display_lines[display_lines.len() - 1us];
- let s = format!("{}:{} ", fm.name, last_line + 1us);
+ let last_line = display_lines[display_lines.len() - 1];
+ let s = format!("{}:{} ", fm.name, last_line + 1);
try!(write!(&mut err.dst, "{0:1$}...\n", "", s.len()));
}
// FIXME (#3260)
// If there's one line at fault we can easily point to the problem
- if lines.lines.len() == 1us {
+ if lines.lines.len() == 1 {
let lo = cm.lookup_char_pos(sp.lo);
- let mut digits = 0us;
- let mut num = (lines.lines[0] + 1us) / 10us;
+ let mut digits = 0;
+ let mut num = (lines.lines[0] + 1) / 10;
// how many digits must be indent past?
- while num > 0us { num /= 10us; digits += 1us; }
+ while num > 0 { num /= 10; digits += 1; }
// indent past |name:## | and the 0-offset column location
- let left = fm.name.len() + digits + lo.col.to_usize() + 3us;
+ let left = fm.name.len() + digits + lo.col.to_usize() + 3;
let mut s = String::new();
// Skip is the number of characters we need to skip because they are
// part of the 'filename:line ' part of the previous line.
- let skip = fm.name.len() + digits + 3us;
+ let skip = fm.name.len() + digits + 3;
for _ in 0..skip {
s.push(' ');
}
if let Some(orig) = fm.get_line(lines.lines[0]) {
- for pos in 0us..left - skip {
+ for pos in 0..left - skip {
let cur_char = orig.as_bytes()[pos] as char;
// Whenever a tab occurs on the previous line, we insert one on
// the error-point-squiggly-line as well (instead of a space).
let hi = cm.lookup_char_pos(sp.hi);
if hi.col != lo.col {
// the ^ already takes up one space
- let num_squigglies = hi.col.to_usize() - lo.col.to_usize() - 1us;
+ let num_squigglies = hi.col.to_usize() - lo.col.to_usize() - 1;
for _ in 0..num_squigglies {
s.push('~');
}
let mut raw_fields = Vec::new(); // ~[[fields of self],
// [fields of next Self arg], [etc]]
let mut patterns = Vec::new();
- for i in 0us..self_args.len() {
+ for i in 0..self_args.len() {
let struct_path= cx.path(DUMMY_SP, vec!( type_ident ));
let (pat, ident_expr) =
trait_.create_struct_pattern(cx,
/// (&A2(ref __self_0),
/// &A2(ref __arg_1_0)) => (*__self_0).eq(&(*__arg_1_0)),
/// _ => {
- /// let __self_vi = match *self { A1(..) => 0us, A2(..) => 1us };
- /// let __arg_1_vi = match *__arg_1 { A1(..) => 0us, A2(..) => 1us };
+ /// let __self_vi = match *self { A1(..) => 0, A2(..) => 1 };
+ /// let __arg_1_vi = match *__arg_1 { A1(..) => 0, A2(..) => 1 };
/// false
/// }
/// }
/// (Variant2, Variant2, Variant2) => ... // delegate Matching on Variant2
/// ...
/// _ => {
- /// let __this_vi = match this { Variant1 => 0us, Variant2 => 1us, ... };
- /// let __that_vi = match that { Variant1 => 0us, Variant2 => 1us, ... };
+ /// let __this_vi = match this { Variant1 => 0, Variant2 => 1, ... };
+ /// let __that_vi = match that { Variant1 => 0, Variant2 => 1, ... };
/// ... // catch-all remainder can inspect above variant index values.
/// }
/// }
//
// ```
// let __self0_vi = match self {
- // A => 0us, B(..) => 1us, C(..) => 2us
+ // A => 0, B(..) => 1, C(..) => 2
// };
// let __self1_vi = match __arg1 {
- // A => 0us, B(..) => 1us, C(..) => 2us
+ // A => 0, B(..) => 1, C(..) => 2
// };
// let __self2_vi = match __arg2 {
- // A => 0us, B(..) => 1us, C(..) => 2us
+ // A => 0, B(..) => 1, C(..) => 2
// };
// ```
let mut index_let_stmts: Vec<P<ast::Stmt>> = Vec::new();
// in this file.
// Token-tree macros:
MacInvocTT(pth, tts, _) => {
- if pth.segments.len() > 1us {
+ if pth.segments.len() > 1 {
fld.cx.span_err(pth.span,
"expected macro name without module \
separators");
},
_ => unreachable!()
};
- if pth.segments.len() > 1us {
+ if pth.segments.len() > 1 {
fld.cx.span_err(pth.span, "expected macro name without module separators");
return DummyResult::raw_pat(span);
}
// try removing it when enough of them are gone.
let mut p = cx.new_parser_from_tts(tts);
- p.quote_depth += 1us;
+ p.quote_depth += 1;
let cx_expr = p.parse_expr();
if !p.eat(&token::Comma) {
stack: vec![],
top_elts: TtSeq(ms),
sep: sep,
- idx: 0us,
+ idx: 0,
up: None,
matches: matches,
- match_lo: 0us,
- match_cur: 0us,
+ match_lo: 0,
+ match_cur: 0,
match_hi: match_idx_hi,
sp_lo: lo
}
}
}
let mut ret_val = HashMap::new();
- let mut idx = 0us;
+ let mut idx = 0;
for m in ms.iter() { n_rec(p_s, m, res, &mut ret_val, &mut idx) }
ret_val
}
if seq.op == ast::ZeroOrMore {
let mut new_ei = ei.clone();
new_ei.match_cur += seq.num_captures;
- new_ei.idx += 1us;
+ new_ei.idx += 1;
//we specifically matched zero repeats.
for idx in ei.match_cur..ei.match_cur + seq.num_captures {
(&mut new_ei.matches[idx]).push(Rc::new(MatchedSeq(vec![], sp)));
cur_eis.push(box MatcherPos {
stack: vec![],
sep: seq.separator.clone(),
- idx: 0us,
+ idx: 0,
matches: matches,
match_lo: ei_t.match_cur,
match_cur: ei_t.match_cur,
/* error messages here could be improved with links to orig. rules */
if token_name_eq(&tok, &token::Eof) {
- if eof_eis.len() == 1us {
+ if eof_eis.len() == 1 {
let mut v = Vec::new();
for dv in (&mut eof_eis[0]).matches.iter_mut() {
v.push(dv.pop().unwrap());
}
return Success(nameize(sess, ms, &v[]));
- } else if eof_eis.len() > 1us {
+ } else if eof_eis.len() > 1 {
return Error(sp, "ambiguity: multiple successful parses".to_string());
} else {
return Failure(sp, "unexpected end of macro invocation".to_string());
}
} else {
- if (bb_eis.len() > 0us && next_eis.len() > 0us)
- || bb_eis.len() > 1us {
+ if (bb_eis.len() > 0 && next_eis.len() > 0)
+ || bb_eis.len() > 1 {
let nts = bb_eis.iter().map(|ei| {
match ei.top_elts.get_tt(ei.idx) {
TtToken(_, MatchNt(bind, name, _, _)) => {
"local ambiguity: multiple parsing options: \
built-in NTs {} or {} other options.",
nts, next_eis.len()).to_string());
- } else if bb_eis.len() == 0us && next_eis.len() == 0us {
+ } else if bb_eis.len() == 0 && next_eis.len() == 0 {
return Failure(sp, format!("no rules expected the token `{}`",
pprust::token_to_string(&tok)).to_string());
- } else if next_eis.len() > 0us {
+ } else if next_eis.len() > 0 {
/* Now process the next token */
- while next_eis.len() > 0us {
+ while next_eis.len() > 0 {
cur_eis.push(next_eis.pop().unwrap());
}
rdr.next_token();
let match_cur = ei.match_cur;
(&mut ei.matches[match_cur]).push(Rc::new(MatchedNonterminal(
parse_nt(&mut rust_parser, span, name_string.get()))));
- ei.idx += 1us;
+ ei.idx += 1;
ei.match_cur += 1;
}
_ => panic!()
}
}
- assert!(cur_eis.len() > 0us);
+ assert!(cur_eis.len() > 0);
}
}
pub fn parse_nt(p: &mut Parser, sp: Span, name: &str) -> Nonterminal {
match name {
"tt" => {
- p.quote_depth += 1us; //but in theory, non-quoted tts might be useful
+ p.quote_depth += 1; //but in theory, non-quoted tts might be useful
let res = token::NtTT(P(p.parse_token_tree()));
- p.quote_depth -= 1us;
+ p.quote_depth -= 1;
return res;
}
_ => {}
r.repeat_len.pop();
}
} else { /* repeat */
- *r.repeat_idx.last_mut().unwrap() += 1us;
+ *r.repeat_idx.last_mut().unwrap() += 1;
r.stack.last_mut().unwrap().idx = 0;
match r.stack.last().unwrap().sep.clone() {
Some(tk) => {
pub fn strip_doc_comment_decoration(comment: &str) -> String {
/// remove whitespace-only lines from the start/end of lines
fn vertical_trim(lines: Vec<String> ) -> Vec<String> {
- let mut i = 0us;
+ let mut i = 0;
let mut j = lines.len();
// first line of all-stars should be omitted
if lines.len() > 0 &&
fn consume_whitespace_counting_blank_lines(rdr: &mut StringReader,
comments: &mut Vec<Comment>) {
while is_whitespace(rdr.curr) && !rdr.is_eof() {
- if rdr.col == CharPos(0us) && rdr.curr_is('\n') {
+ if rdr.col == CharPos(0) && rdr.curr_is('\n') {
push_blank_line_comment(rdr, &mut *comments);
}
rdr.bump();
let mut style = if code_to_the_left { Trailing } else { Isolated };
rdr.consume_non_eol_whitespace();
- if !rdr.is_eof() && !rdr.curr_is('\n') && lines.len() == 1us {
+ if !rdr.is_eof() && !rdr.curr_is('\n') && lines.len() == 1 {
style = Mixed;
}
debug!("<<< block comment");
/// Converts CRLF to LF in the given string, raising an error on bare CR.
fn translate_crlf<'b>(&self, start: BytePos,
s: &'b str, errmsg: &'b str) -> CowString<'b> {
- let mut i = 0us;
+ let mut i = 0;
while i < s.len() {
let str::CharRange { ch, next } = s.char_range_at(i);
if ch == '\r' {
let byte_offset_diff = next.next - current_byte_offset;
self.pos = self.pos + Pos::from_usize(byte_offset_diff);
self.curr = Some(next.ch);
- self.col = self.col + CharPos(1us);
+ self.col = self.col + CharPos(1);
if last_char == '\n' {
self.filemap.next_line(self.last_pos);
- self.col = CharPos(0us);
+ self.col = CharPos(0);
}
if byte_offset_diff > 1 {
cmap.files.borrow_mut().push(self.filemap.clone());
let loc = cmap.lookup_char_pos_adj(self.last_pos);
debug!("Skipping a shebang");
- if loc.line == 1us && loc.col == CharPos(0us) {
+ if loc.line == 1 && loc.col == CharPos(0) {
// FIXME: Add shebang "token", return it
let start = self.last_pos;
while !self.curr_is('\n') && !self.is_eof() { self.bump(); }
/// Scan through any digits (base `radix`) or underscores, and return how
/// many digits there were.
fn scan_digits(&mut self, radix: usize) -> usize {
- let mut len = 0us;
+ let mut len = 0;
loop {
let c = self.curr;
if c == Some('_') { debug!("skipping a _"); self.bump(); continue; }
if self.curr == Some('{') {
self.scan_unicode_escape(delim)
} else {
- let res = self.scan_hex_digits(4us, delim, false);
+ let res = self.scan_hex_digits(4, delim, false);
let sp = codemap::mk_sp(escaped_pos, self.last_pos);
self.old_escape_warning(sp);
res
}
}
'U' if !ascii_only => {
- let res = self.scan_hex_digits(8us, delim, false);
+ let res = self.scan_hex_digits(8, delim, false);
let sp = codemap::mk_sp(escaped_pos, self.last_pos);
self.old_escape_warning(sp);
res
fn scan_unicode_escape(&mut self, delim: char) -> bool {
self.bump(); // past the {
let start_bpos = self.last_pos;
- let mut count = 0us;
+ let mut count = 0;
let mut accum_int = 0;
while !self.curr_is('}') && count <= 6 {
/// error if it isn't.
fn check_float_base(&mut self, start_bpos: BytePos, last_bpos: BytePos, base: usize) {
match base {
- 16us => self.err_span_(start_bpos, last_bpos, "hexadecimal float literal is not \
+ 16 => self.err_span_(start_bpos, last_bpos, "hexadecimal float literal is not \
supported"),
- 8us => self.err_span_(start_bpos, last_bpos, "octal float literal is not supported"),
- 2us => self.err_span_(start_bpos, last_bpos, "binary float literal is not supported"),
+ 8 => self.err_span_(start_bpos, last_bpos, "octal float literal is not supported"),
+ 2 => self.err_span_(start_bpos, last_bpos, "binary float literal is not supported"),
_ => ()
}
}
'r' => {
let start_bpos = self.last_pos;
self.bump();
- let mut hash_count = 0us;
+ let mut hash_count = 0;
while self.curr_is('#') {
self.bump();
hash_count += 1;
fn scan_raw_byte_string(&mut self) -> token::Lit {
let start_bpos = self.last_pos;
self.bump();
- let mut hash_count = 0us;
+ let mut hash_count = 0;
while self.curr_is('#') {
self.bump();
hash_count += 1;
name,
source
);
- p.quote_depth += 1us;
+ p.quote_depth += 1;
// right now this is re-creating the token trees from ... token trees.
maybe_aborted(p.parse_all_token_trees(),p)
}
name,
source
);
- p.quote_depth += 1us;
+ p.quote_depth += 1;
// right now this is re-creating the token trees from ... token trees.
maybe_aborted(p.parse_all_token_trees(),p)
}
match suffix {
Some(suf) if looks_like_width_suffix(&['f'], suf) => {
match base {
- 16us => sd.span_err(sp, "hexadecimal float literal is not supported"),
- 8us => sd.span_err(sp, "octal float literal is not supported"),
- 2us => sd.span_err(sp, "binary float literal is not supported"),
+ 16 => sd.span_err(sp, "hexadecimal float literal is not supported"),
+ 8 => sd.span_err(sp, "octal float literal is not supported"),
+ 2 => sd.span_err(sp, "binary float literal is not supported"),
_ => ()
}
let ident = token::intern_and_get_ident(&*s);
// would encounter a `>` and stop. This lets the parser handle trailing
// commas in generic parameters, because it can stop either after
// parsing a type or after parsing a comma.
- for i in iter::count(0us, 1) {
+ for i in iter::count(0, 1) {
if self.check(&token::Gt)
|| self.token == token::BinOp(token::Shr)
|| self.token == token::Ge
};
self.span = next.sp;
self.token = next.tok;
- self.tokens_consumed += 1us;
+ self.tokens_consumed += 1;
self.expected_tokens.clear();
// check after each token
self.check_unknown_macro_variable();
}
pub fn check_unknown_macro_variable(&mut self) {
- if self.quote_depth == 0us {
+ if self.quote_depth == 0 {
match self.token {
token::SubstNt(name, _) =>
self.fatal(&format!("unknown macro variable `{}`",
token_str)[])
},
/* we ought to allow different depths of unquotation */
- token::Dollar | token::SubstNt(..) if p.quote_depth > 0us => {
+ token::Dollar | token::SubstNt(..) if p.quote_depth > 0 => {
p.parse_unquoted()
}
_ => {
return Ok(item);
}
if self.token.is_keyword(keywords::Unsafe) &&
- self.look_ahead(1us, |t| t.is_keyword(keywords::Trait))
+ self.look_ahead(1, |t| t.is_keyword(keywords::Trait))
{
// UNSAFE TRAIT ITEM
self.expect_keyword(keywords::Unsafe);
return Ok(item);
}
if self.token.is_keyword(keywords::Unsafe) &&
- self.look_ahead(1us, |t| t.is_keyword(keywords::Impl))
+ self.look_ahead(1, |t| t.is_keyword(keywords::Impl))
{
// IMPL ITEM
self.expect_keyword(keywords::Unsafe);
return Ok(item);
}
if self.token.is_keyword(keywords::Unsafe)
- && self.look_ahead(1us, |t| *t != token::OpenDelim(token::Brace)) {
+ && self.look_ahead(1, |t| *t != token::OpenDelim(token::Brace)) {
// UNSAFE FUNCTION ITEM
self.bump();
let abi = if self.eat_keyword(keywords::Extern) {
}
}
}
- let mut rename_to = path[path.len() - 1us];
+ let mut rename_to = path[path.len() - 1];
let path = ast::Path {
span: mk_sp(lo, self.last_span.hi),
global: false,
let mut i = left;
let mut l = lim;
let mut s = string::String::from_str("[");
- while i != right && l != 0us {
- l -= 1us;
+ while i != right && l != 0 {
+ l -= 1;
if i != left {
s.push_str(", ");
}
s.push_str(&format!("{}={}",
szs[i],
tok_str(&toks[i]))[]);
- i += 1us;
+ i += 1;
i %= n;
}
s.push(']');
if self.scan_stack_empty {
self.left_total = 1;
self.right_total = 1;
- self.left = 0us;
- self.right = 0us;
+ self.left = 0;
+ self.right = 0;
} else { self.advance_right(); }
debug!("pp Begin({})/buffer ~[{},{}]",
b.offset, self.left, self.right);
if self.scan_stack_empty {
self.left_total = 1;
self.right_total = 1;
- self.left = 0us;
- self.right = 0us;
+ self.left = 0;
+ self.right = 0;
} else { self.advance_right(); }
debug!("pp Break({})/buffer ~[{},{}]",
b.offset, self.left, self.right);
if self.scan_stack_empty {
self.scan_stack_empty = false;
} else {
- self.top += 1us;
+ self.top += 1;
self.top %= self.buf_len;
assert!((self.top != self.bottom));
}
if self.top == self.bottom {
self.scan_stack_empty = true;
} else {
- self.top += self.buf_len - 1us; self.top %= self.buf_len;
+ self.top += self.buf_len - 1; self.top %= self.buf_len;
}
return x;
}
if self.top == self.bottom {
self.scan_stack_empty = true;
} else {
- self.bottom += 1us; self.bottom %= self.buf_len;
+ self.bottom += 1; self.bottom %= self.buf_len;
}
return x;
}
pub fn advance_right(&mut self) {
- self.right += 1us;
+ self.right += 1;
self.right %= self.buf_len;
assert!((self.right != self.left));
}
break;
}
- self.left += 1us;
+ self.left += 1;
self.left %= self.buf_len;
left_size = self.size[self.left];
pub fn get_top(&mut self) -> PrintStackElem {
let print_stack = &mut self.print_stack;
let n = print_stack.len();
- if n != 0us {
+ if n != 0 {
(*print_stack)[n - 1]
} else {
PrintStackElem {
Token::End => {
debug!("print End -> pop End");
let print_stack = &mut self.print_stack;
- assert!((print_stack.len() != 0us));
+ assert!((print_stack.len() != 0));
print_stack.pop().unwrap();
Ok(())
}
}
pub fn zerobreak(p: &mut Printer) -> old_io::IoResult<()> {
- spaces(p, 0us)
+ spaces(p, 0)
}
pub fn space(p: &mut Printer) -> old_io::IoResult<()> {
- spaces(p, 1us)
+ spaces(p, 1)
}
pub fn hardbreak(p: &mut Printer) -> old_io::IoResult<()> {
}
#[allow(non_upper_case_globals)]
-pub const indent_unit: usize = 4us;
+pub const indent_unit: usize = 4;
#[allow(non_upper_case_globals)]
-pub const default_columns: usize = 78us;
+pub const default_columns: usize = 78;
/// Requires you to pass an input filename and reader so that
/// it can scan the input text for comments and literals to
// containing cbox, will be closed by print-block at }
try!(s.cbox(indent_unit));
// head-ibox, will be closed by print-block after {
- try!(s.ibox(0us));
+ try!(s.ibox(0));
s.print_block(blk)
})
}
pub fn bclose_maybe_open (&mut self, span: codemap::Span,
indented: usize, close_box: bool) -> IoResult<()> {
try!(self.maybe_print_comment(span.hi));
- try!(self.break_offset_if_not_bol(1us, -(indented as isize)));
+ try!(self.break_offset_if_not_bol(1, -(indented as isize)));
try!(word(&mut self.s, "}"));
if close_box {
try!(self.end()); // close the outer-box
pub fn commasep<T, F>(&mut self, b: Breaks, elts: &[T], mut op: F) -> IoResult<()> where
F: FnMut(&mut State, &T) -> IoResult<()>,
{
- try!(self.rbox(0us, b));
+ try!(self.rbox(0, b));
let mut first = true;
for elt in elts.iter() {
if first { first = false; } else { try!(self.word_space(",")); }
F: FnMut(&mut State, &T) -> IoResult<()>,
G: FnMut(&T) -> codemap::Span,
{
- try!(self.rbox(0us, b));
+ try!(self.rbox(0, b));
let len = elts.len();
- let mut i = 0us;
+ let mut i = 0;
for elt in elts.iter() {
try!(self.maybe_print_comment(get_span(elt).hi));
try!(op(self, elt));
- i += 1us;
+ i += 1;
if i < len {
try!(word(&mut self.s, ","));
try!(self.maybe_print_trailing_comment(get_span(elt),
pub fn print_type(&mut self, ty: &ast::Ty) -> IoResult<()> {
try!(self.maybe_print_comment(ty.span.lo));
- try!(self.ibox(0us));
+ try!(self.ibox(0));
match ty.node {
ast::TyVec(ref ty) => {
try!(word(&mut self.s, "["));
}
ast::ItemTy(ref ty, ref params) => {
try!(self.ibox(indent_unit));
- try!(self.ibox(0us));
+ try!(self.ibox(0));
try!(self.word_nbsp(&visibility_qualified(item.vis, "type")[]));
try!(self.print_ident(item.ident));
try!(self.print_generics(params));
pub fn print_outer_attributes(&mut self,
attrs: &[ast::Attribute]) -> IoResult<()> {
- let mut count = 0us;
+ let mut count = 0;
for attr in attrs.iter() {
match attr.node.style {
ast::AttrOuter => {
pub fn print_inner_attributes(&mut self,
attrs: &[ast::Attribute]) -> IoResult<()> {
- let mut count = 0us;
+ let mut count = 0;
for attr in attrs.iter() {
match attr.node.style {
ast::AttrInner => {
match _else.node {
// "another else-if"
ast::ExprIf(ref i, ref then, ref e) => {
- try!(self.cbox(indent_unit - 1us));
- try!(self.ibox(0us));
+ try!(self.cbox(indent_unit - 1));
+ try!(self.ibox(0));
try!(word(&mut self.s, " else if "));
try!(self.print_expr(&**i));
try!(space(&mut self.s));
}
// "another else-if-let"
ast::ExprIfLet(ref pat, ref expr, ref then, ref e) => {
- try!(self.cbox(indent_unit - 1us));
- try!(self.ibox(0us));
+ try!(self.cbox(indent_unit - 1));
+ try!(self.ibox(0));
try!(word(&mut self.s, " else if let "));
try!(self.print_pat(&**pat));
try!(space(&mut self.s));
}
// "final else"
ast::ExprBlock(ref b) => {
- try!(self.cbox(indent_unit - 1us));
- try!(self.ibox(0us));
+ try!(self.cbox(indent_unit - 1));
+ try!(self.ibox(0));
try!(word(&mut self.s, " else "));
self.print_block(&**b)
}
try!(self.print_expr(&*args[0]));
try!(word(&mut self.s, "."));
try!(self.print_ident(ident.node));
- if tys.len() > 0us {
+ if tys.len() > 0 {
try!(word(&mut self.s, "::<"));
try!(self.commasep(Inconsistent, tys,
|s, ty| s.print_type(&**ty)));
// containing cbox, will be closed by print-block at }
try!(self.cbox(indent_unit));
// head-box, will be closed by print-block after {
- try!(self.ibox(0us));
+ try!(self.ibox(0));
try!(self.print_block(&**blk));
}
ast::ExprAssign(ref lhs, ref rhs) => {
},
|f| f.node.pat.span));
if etc {
- if fields.len() != 0us { try!(self.word_space(",")); }
+ if fields.len() != 0 { try!(self.word_space(",")); }
try!(word(&mut self.s, ".."));
}
try!(space(&mut self.s));
try!(space(&mut self.s));
}
try!(self.cbox(indent_unit));
- try!(self.ibox(0us));
+ try!(self.ibox(0));
try!(self.print_outer_attributes(&arm.attrs[]));
let mut first = true;
for p in arm.pats.iter() {
-> IoResult<()> {
// It is unfortunate to duplicate the commasep logic, but we want the
// self type and the args all in the same box.
- try!(self.rbox(0us, Inconsistent));
+ try!(self.rbox(0, Inconsistent));
let mut first = true;
for &explicit_self in opt_explicit_self.iter() {
let m = match explicit_self {
try!(word(&mut self.s, "<"));
let mut ints = Vec::new();
- for i in 0us..total {
+ for i in 0..total {
ints.push(i);
}
if span.hi < (*cmnt).pos && (*cmnt).pos < next &&
span_line.line == comment_line.line {
try!(self.print_comment(cmnt));
- self.cur_cmnt_and_lit.cur_cmnt += 1us;
+ self.cur_cmnt_and_lit.cur_cmnt += 1;
}
}
_ => ()
match self.next_comment() {
Some(ref cmnt) => {
try!(self.print_comment(cmnt));
- self.cur_cmnt_and_lit.cur_cmnt += 1us;
+ self.cur_cmnt_and_lit.cur_cmnt += 1;
}
_ => break
}
while self.cur_cmnt_and_lit.cur_lit < lits.len() {
let ltrl = (*lits)[self.cur_cmnt_and_lit.cur_lit].clone();
if ltrl.pos > pos { return None; }
- self.cur_cmnt_and_lit.cur_lit += 1us;
+ self.cur_cmnt_and_lit.cur_lit += 1;
if ltrl.pos == pos { return Some(ltrl); }
}
None
Some(ref cmnt) => {
if (*cmnt).pos < pos {
try!(self.print_comment(cmnt));
- self.cur_cmnt_and_lit.cur_cmnt += 1us;
+ self.cur_cmnt_and_lit.cur_cmnt += 1;
} else { break; }
}
_ => break
cmnt: &comments::Comment) -> IoResult<()> {
match cmnt.style {
comments::Mixed => {
- assert_eq!(cmnt.lines.len(), 1us);
+ assert_eq!(cmnt.lines.len(), 1);
try!(zerobreak(&mut self.s));
try!(word(&mut self.s, &cmnt.lines[0][]));
zerobreak(&mut self.s)
}
comments::Trailing => {
try!(word(&mut self.s, " "));
- if cmnt.lines.len() == 1us {
+ if cmnt.lines.len() == 1 {
try!(word(&mut self.s, &cmnt.lines[0][]));
hardbreak(&mut self.s)
} else {
- try!(self.ibox(0us));
+ try!(self.ibox(0));
for line in cmnt.lines.iter() {
if !line.is_empty() {
try!(word(&mut self.s, &line[]));
let tparm_cnt = generics.ty_params.len();
// NB: inadequate check, but we're running
// well before resolve, can't get too deep.
- input_cnt == 1us
- && no_output && tparm_cnt == 0us
+ input_cnt == 1
+ && no_output && tparm_cnt == 0
}
_ => false
}