Replace full slice notation with index calls

This commit is contained in:
Nick Cameron 2015-01-02 13:56:28 +13:00
parent 918255ef8c
commit f7ff37e4c5
225 changed files with 2166 additions and 2139 deletions

View file

@ -100,7 +100,7 @@ pub fn expand_asm<'cx>(cx: &'cx mut ExtCtxt, sp: Span, tts: &[ast::TokenTree])
Some(('=', _)) => None,
Some(('+', operand)) => {
Some(token::intern_and_get_ident(format!(
"={}", operand)[]))
"={}", operand).index(&FullRange)))
}
_ => {
cx.span_err(span, "output operand constraint lacks '=' or '+'");

View file

@ -539,7 +539,7 @@ impl<'a> ExtCtxt<'a> {
pub fn mod_pop(&mut self) { self.mod_path.pop().unwrap(); }
pub fn mod_path(&self) -> Vec<ast::Ident> {
let mut v = Vec::new();
v.push(token::str_to_ident(self.ecfg.crate_name[]));
v.push(token::str_to_ident(self.ecfg.crate_name.index(&FullRange)));
v.extend(self.mod_path.iter().map(|a| *a));
return v;
}
@ -548,7 +548,7 @@ impl<'a> ExtCtxt<'a> {
if self.recursion_count > self.ecfg.recursion_limit {
self.span_fatal(ei.call_site,
format!("recursion limit reached while expanding the macro `{}`",
ei.callee.name)[]);
ei.callee.name).index(&FullRange));
}
let mut call_site = ei.call_site;
@ -670,7 +670,7 @@ pub fn check_zero_tts(cx: &ExtCtxt,
tts: &[ast::TokenTree],
name: &str) {
if tts.len() != 0 {
cx.span_err(sp, format!("{} takes no arguments", name)[]);
cx.span_err(sp, format!("{} takes no arguments", name).index(&FullRange));
}
}
@ -683,12 +683,12 @@ pub fn get_single_str_from_tts(cx: &mut ExtCtxt,
-> Option<String> {
let mut p = cx.new_parser_from_tts(tts);
if p.token == token::Eof {
cx.span_err(sp, format!("{} takes 1 argument", name)[]);
cx.span_err(sp, format!("{} takes 1 argument", name).index(&FullRange));
return None
}
let ret = cx.expander().fold_expr(p.parse_expr());
if p.token != token::Eof {
cx.span_err(sp, format!("{} takes 1 argument", name)[]);
cx.span_err(sp, format!("{} takes 1 argument", name).index(&FullRange));
}
expr_to_string(cx, ret, "argument must be a string literal").map(|(s, _)| {
s.get().to_string()

View file

@ -709,7 +709,7 @@ impl<'a> AstBuilder for ExtCtxt<'a> {
let loc = self.codemap().lookup_char_pos(span.lo);
let expr_file = self.expr_str(span,
token::intern_and_get_ident(loc.file
.name[]));
.name.index(&FullRange)));
let expr_line = self.expr_uint(span, loc.line);
let expr_file_line_tuple = self.expr_tuple(span, vec!(expr_file, expr_line));
let expr_file_line_ptr = self.expr_addr_of(span, expr_file_line_tuple);

View file

@ -40,14 +40,14 @@ pub fn expand_syntax_ext(cx: &mut base::ExtCtxt,
ast::LitInt(i, ast::UnsignedIntLit(_)) |
ast::LitInt(i, ast::SignedIntLit(_, ast::Plus)) |
ast::LitInt(i, ast::UnsuffixedIntLit(ast::Plus)) => {
accumulator.push_str(format!("{}", i)[]);
accumulator.push_str(format!("{}", i).index(&FullRange));
}
ast::LitInt(i, ast::SignedIntLit(_, ast::Minus)) |
ast::LitInt(i, ast::UnsuffixedIntLit(ast::Minus)) => {
accumulator.push_str(format!("-{}", i)[]);
accumulator.push_str(format!("-{}", i).index(&FullRange));
}
ast::LitBool(b) => {
accumulator.push_str(format!("{}", b)[]);
accumulator.push_str(format!("{}", b).index(&FullRange));
}
ast::LitByte(..) |
ast::LitBinary(..) => {
@ -62,5 +62,5 @@ pub fn expand_syntax_ext(cx: &mut base::ExtCtxt,
}
base::MacExpr::new(cx.expr_str(
sp,
token::intern_and_get_ident(accumulator[])))
token::intern_and_get_ident(accumulator.index(&FullRange))))
}

View file

@ -40,7 +40,7 @@ pub fn expand_syntax_ext<'cx>(cx: &mut ExtCtxt, sp: Span, tts: &[ast::TokenTree]
}
}
}
let res = str_to_ident(res_str[]);
let res = str_to_ident(res_str.index(&FullRange));
let e = P(ast::Expr {
id: ast::DUMMY_NODE_ID,

View file

@ -29,12 +29,12 @@ pub fn expand_deriving_bound<F>(cx: &mut ExtCtxt,
"Send" | "Sync" => {
return cx.span_err(span,
format!("{} is an unsafe trait and it \
should be implemented explicitly", *tname)[])
should be implemented explicitly", *tname).index(&FullRange))
}
ref tname => {
cx.span_bug(span,
format!("expected built-in trait name but \
found {}", *tname)[])
found {}", *tname).index(&FullRange))
}
}
},

View file

@ -80,11 +80,11 @@ fn cs_clone(
EnumNonMatchingCollapsed (..) => {
cx.span_bug(trait_span,
format!("non-matching enum variants in \
`deriving({})`", name)[])
`deriving({})`", name).index(&FullRange))
}
StaticEnum(..) | StaticStruct(..) => {
cx.span_bug(trait_span,
format!("static method in `deriving({})`", name)[])
format!("static method in `deriving({})`", name).index(&FullRange))
}
}
@ -101,7 +101,7 @@ fn cs_clone(
None => {
cx.span_bug(trait_span,
format!("unnamed field in normal struct in \
`deriving({})`", name)[])
`deriving({})`", name).index(&FullRange))
}
};
cx.field_imm(field.span, ident, subcall(field))

View file

@ -198,7 +198,7 @@ fn decode_static_fields<F>(cx: &mut ExtCtxt,
let fields = fields.iter().enumerate().map(|(i, &span)| {
getarg(cx, span,
token::intern_and_get_ident(format!("_field{}",
i)[]),
i).index(&FullRange)),
i)
}).collect();

View file

@ -183,7 +183,7 @@ fn encodable_substructure(cx: &mut ExtCtxt, trait_span: Span,
let name = match name {
Some(id) => token::get_ident(id),
None => {
token::intern_and_get_ident(format!("_field{}", i)[])
token::intern_and_get_ident(format!("_field{}", i).index(&FullRange))
}
};
let enc = cx.expr_method_call(span, self_.clone(),

View file

@ -510,15 +510,15 @@ impl<'a> TraitDef<'a> {
self,
struct_def,
type_ident,
self_args[],
nonself_args[])
self_args.index(&FullRange),
nonself_args.index(&FullRange))
} else {
method_def.expand_struct_method_body(cx,
self,
struct_def,
type_ident,
self_args[],
nonself_args[])
self_args.index(&FullRange),
nonself_args.index(&FullRange))
};
method_def.create_method(cx,
@ -550,15 +550,15 @@ impl<'a> TraitDef<'a> {
self,
enum_def,
type_ident,
self_args[],
nonself_args[])
self_args.index(&FullRange),
nonself_args.index(&FullRange))
} else {
method_def.expand_enum_method_body(cx,
self,
enum_def,
type_ident,
self_args,
nonself_args[])
nonself_args.index(&FullRange))
};
method_def.create_method(cx,
@ -645,7 +645,7 @@ impl<'a> MethodDef<'a> {
for (i, ty) in self.args.iter().enumerate() {
let ast_ty = ty.to_ty(cx, trait_.span, type_ident, generics);
let ident = cx.ident_of(format!("__arg_{}", i)[]);
let ident = cx.ident_of(format!("__arg_{}", i).index(&FullRange));
arg_tys.push((ident, ast_ty));
let arg_expr = cx.expr_ident(trait_.span, ident);
@ -752,7 +752,7 @@ impl<'a> MethodDef<'a> {
struct_path,
struct_def,
format!("__self_{}",
i)[],
i).index(&FullRange),
ast::MutImmutable);
patterns.push(pat);
raw_fields.push(ident_expr);
@ -908,22 +908,22 @@ impl<'a> MethodDef<'a> {
.collect::<Vec<String>>();
let self_arg_idents = self_arg_names.iter()
.map(|name|cx.ident_of(name[]))
.map(|name|cx.ident_of(name.index(&FullRange)))
.collect::<Vec<ast::Ident>>();
// The `vi_idents` will be bound, solely in the catch-all, to
// a series of let statements mapping each self_arg to a uint
// corresponding to its variant index.
let vi_idents: Vec<ast::Ident> = self_arg_names.iter()
.map(|name| { let vi_suffix = format!("{}_vi", name[]);
cx.ident_of(vi_suffix[]) })
.map(|name| { let vi_suffix = format!("{}_vi", name.index(&FullRange));
cx.ident_of(vi_suffix.index(&FullRange)) })
.collect::<Vec<ast::Ident>>();
// Builds, via callback to call_substructure_method, the
// delegated expression that handles the catch-all case,
// using `__variants_tuple` to drive logic if necessary.
let catch_all_substructure = EnumNonMatchingCollapsed(
self_arg_idents, variants[], vi_idents[]);
self_arg_idents, variants.index(&FullRange), vi_idents.index(&FullRange));
// These arms are of the form:
// (Variant1, Variant1, ...) => Body1
@ -945,12 +945,12 @@ impl<'a> MethodDef<'a> {
let mut subpats = Vec::with_capacity(self_arg_names.len());
let mut self_pats_idents = Vec::with_capacity(self_arg_names.len() - 1);
let first_self_pat_idents = {
let (p, idents) = mk_self_pat(cx, self_arg_names[0][]);
let (p, idents) = mk_self_pat(cx, self_arg_names[0].index(&FullRange));
subpats.push(p);
idents
};
for self_arg_name in self_arg_names.tail().iter() {
let (p, idents) = mk_self_pat(cx, self_arg_name[]);
let (p, idents) = mk_self_pat(cx, self_arg_name.index(&FullRange));
subpats.push(p);
self_pats_idents.push(idents);
}
@ -1006,7 +1006,7 @@ impl<'a> MethodDef<'a> {
&**variant,
field_tuples);
let arm_expr = self.call_substructure_method(
cx, trait_, type_ident, self_args[], nonself_args,
cx, trait_, type_ident, self_args.index(&FullRange), nonself_args,
&substructure);
cx.arm(sp, vec![single_pat], arm_expr)
@ -1059,7 +1059,7 @@ impl<'a> MethodDef<'a> {
}
let arm_expr = self.call_substructure_method(
cx, trait_, type_ident, self_args[], nonself_args,
cx, trait_, type_ident, self_args.index(&FullRange), nonself_args,
&catch_all_substructure);
// Builds the expression:
@ -1263,7 +1263,7 @@ impl<'a> TraitDef<'a> {
cx.span_bug(sp, "a struct with named and unnamed fields in `derive`");
}
};
let ident = cx.ident_of(format!("{}_{}", prefix, i)[]);
let ident = cx.ident_of(format!("{}_{}", prefix, i).index(&FullRange));
paths.push(codemap::Spanned{span: sp, node: ident});
let val = cx.expr(
sp, ast::ExprParen(cx.expr_deref(sp, cx.expr_path(cx.path_ident(sp,ident)))));
@ -1309,7 +1309,7 @@ impl<'a> TraitDef<'a> {
let mut ident_expr = Vec::new();
for (i, va) in variant_args.iter().enumerate() {
let sp = self.set_expn_info(cx, va.ty.span);
let ident = cx.ident_of(format!("{}_{}", prefix, i)[]);
let ident = cx.ident_of(format!("{}_{}", prefix, i).index(&FullRange));
let path1 = codemap::Spanned{span: sp, node: ident};
paths.push(path1);
let expr_path = cx.expr_path(cx.path_ident(sp, ident));
@ -1352,7 +1352,7 @@ pub fn cs_fold<F>(use_foldl: bool,
field.span,
old,
field.self_.clone(),
field.other[])
field.other.index(&FullRange))
})
} else {
all_fields.iter().rev().fold(base, |old, field| {
@ -1360,12 +1360,12 @@ pub fn cs_fold<F>(use_foldl: bool,
field.span,
old,
field.self_.clone(),
field.other[])
field.other.index(&FullRange))
})
}
},
EnumNonMatchingCollapsed(ref all_args, _, tuple) =>
enum_nonmatch_f.call_mut((cx, trait_span, (all_args[], tuple),
enum_nonmatch_f.call_mut((cx, trait_span, (all_args.index(&FullRange), tuple),
substructure.nonself_args)),
StaticEnum(..) | StaticStruct(..) => {
cx.span_bug(trait_span, "static function in `derive`")
@ -1405,7 +1405,7 @@ pub fn cs_same_method<F>(f: F,
f(cx, trait_span, called)
},
EnumNonMatchingCollapsed(ref all_self_args, _, tuple) =>
enum_nonmatch_f.call_mut((cx, trait_span, (all_self_args[], tuple),
enum_nonmatch_f.call_mut((cx, trait_span, (all_self_args.index(&FullRange), tuple),
substructure.nonself_args)),
StaticEnum(..) | StaticStruct(..) => {
cx.span_bug(trait_span, "static function in `derive`")

View file

@ -123,7 +123,7 @@ pub fn expand_meta_derive(cx: &mut ExtCtxt,
cx.span_err(titem.span,
format!("unknown `derive` \
trait: `{}`",
*tname)[]);
*tname).index(&FullRange));
}
};
}

View file

@ -127,7 +127,7 @@ fn show_substructure(cx: &mut ExtCtxt, span: Span,
let formatter = substr.nonself_args[0].clone();
let meth = cx.ident_of("write_fmt");
let s = token::intern_and_get_ident(format_string[]);
let s = token::intern_and_get_ident(format_string.index(&FullRange));
let format_string = cx.expr_str(span, s);
// phew, not our responsibility any more!

View file

@ -30,7 +30,7 @@ pub fn expand_option_env<'cx>(cx: &'cx mut ExtCtxt, sp: Span, tts: &[ast::TokenT
Some(v) => v
};
let e = match os::getenv(var[]) {
let e = match os::getenv(var.index(&FullRange)) {
None => {
cx.expr_path(cx.path_all(sp,
true,
@ -56,7 +56,7 @@ pub fn expand_option_env<'cx>(cx: &'cx mut ExtCtxt, sp: Span, tts: &[ast::TokenT
cx.ident_of("Some")),
vec!(cx.expr_str(sp,
token::intern_and_get_ident(
s[]))))
s.index(&FullRange)))))
}
};
MacExpr::new(e)
@ -83,7 +83,7 @@ pub fn expand_env<'cx>(cx: &'cx mut ExtCtxt, sp: Span, tts: &[ast::TokenTree])
None => {
token::intern_and_get_ident(format!("environment variable `{}` \
not defined",
var)[])
var).index(&FullRange))
}
Some(second) => {
match expr_to_string(cx, second, "expected string literal") {
@ -106,7 +106,7 @@ pub fn expand_env<'cx>(cx: &'cx mut ExtCtxt, sp: Span, tts: &[ast::TokenTree])
cx.span_err(sp, msg.get());
cx.expr_uint(sp, 0)
}
Some(s) => cx.expr_str(sp, token::intern_and_get_ident(s[]))
Some(s) => cx.expr_str(sp, token::intern_and_get_ident(s.index(&FullRange)))
};
MacExpr::new(e)
}

View file

@ -287,7 +287,7 @@ fn expand_mac_invoc<T, F, G>(mac: ast::Mac, span: codemap::Span,
fld.cx.span_err(
pth.span,
format!("macro undefined: '{}!'",
extnamestr.get())[]);
extnamestr.get()).index(&FullRange));
// let compilation continue
None
@ -303,7 +303,7 @@ fn expand_mac_invoc<T, F, G>(mac: ast::Mac, span: codemap::Span,
},
});
let fm = fresh_mark();
let marked_before = mark_tts(tts[], fm);
let marked_before = mark_tts(tts.index(&FullRange), fm);
// The span that we pass to the expanders we want to
// be the root of the call stack. That's the most
@ -314,7 +314,7 @@ fn expand_mac_invoc<T, F, G>(mac: ast::Mac, span: codemap::Span,
let opt_parsed = {
let expanded = expandfun.expand(fld.cx,
mac_span,
marked_before[]);
marked_before.index(&FullRange));
parse_thunk(expanded)
};
let parsed = match opt_parsed {
@ -323,8 +323,8 @@ fn expand_mac_invoc<T, F, G>(mac: ast::Mac, span: codemap::Span,
fld.cx.span_err(
pth.span,
format!("non-expression macro in expression position: {}",
extnamestr.get()[]
)[]);
extnamestr.get().index(&FullRange)
).index(&FullRange));
return None;
}
};
@ -334,7 +334,7 @@ fn expand_mac_invoc<T, F, G>(mac: ast::Mac, span: codemap::Span,
fld.cx.span_err(
pth.span,
format!("'{}' is not a tt-style macro",
extnamestr.get())[]);
extnamestr.get()).index(&FullRange));
None
}
}
@ -439,7 +439,7 @@ pub fn expand_item(it: P<ast::Item>, fld: &mut MacroExpander)
if valid_ident {
fld.cx.mod_push(it.ident);
}
let macro_use = contains_macro_use(fld, new_attrs[]);
let macro_use = contains_macro_use(fld, new_attrs.index(&FullRange));
let result = with_exts_frame!(fld.cx.syntax_env,
macro_use,
noop_fold_item(it, fld));
@ -566,7 +566,7 @@ pub fn expand_item_mac(it: P<ast::Item>,
None => {
fld.cx.span_err(path_span,
format!("macro undefined: '{}!'",
extnamestr)[]);
extnamestr).index(&FullRange));
// let compilation continue
return SmallVector::zero();
}
@ -579,7 +579,7 @@ pub fn expand_item_mac(it: P<ast::Item>,
format!("macro {}! expects no ident argument, \
given '{}'",
extnamestr,
token::get_ident(it.ident))[]);
token::get_ident(it.ident)).index(&FullRange));
return SmallVector::zero();
}
fld.cx.bt_push(ExpnInfo {
@ -591,14 +591,14 @@ pub fn expand_item_mac(it: P<ast::Item>,
}
});
// mark before expansion:
let marked_before = mark_tts(tts[], fm);
expander.expand(fld.cx, it.span, marked_before[])
let marked_before = mark_tts(tts.index(&FullRange), fm);
expander.expand(fld.cx, it.span, marked_before.index(&FullRange))
}
IdentTT(ref expander, span) => {
if it.ident.name == parse::token::special_idents::invalid.name {
fld.cx.span_err(path_span,
format!("macro {}! expects an ident argument",
extnamestr.get())[]);
extnamestr.get()).index(&FullRange));
return SmallVector::zero();
}
fld.cx.bt_push(ExpnInfo {
@ -610,13 +610,14 @@ pub fn expand_item_mac(it: P<ast::Item>,
}
});
// mark before expansion:
let marked_tts = mark_tts(tts[], fm);
let marked_tts = mark_tts(tts.index(&FullRange), fm);
expander.expand(fld.cx, it.span, it.ident, marked_tts)
}
MacroRulesTT => {
if it.ident.name == parse::token::special_idents::invalid.name {
fld.cx.span_err(path_span,
format!("macro_rules! expects an ident argument")[]);
format!("macro_rules! expects an ident argument")
.index(&FullRange));
return SmallVector::zero();
}
fld.cx.bt_push(ExpnInfo {
@ -648,7 +649,7 @@ pub fn expand_item_mac(it: P<ast::Item>,
_ => {
fld.cx.span_err(it.span,
format!("{}! is not legal in item position",
extnamestr.get())[]);
extnamestr.get()).index(&FullRange));
return SmallVector::zero();
}
}
@ -667,7 +668,7 @@ pub fn expand_item_mac(it: P<ast::Item>,
None => {
fld.cx.span_err(path_span,
format!("non-item macro in item position: {}",
extnamestr.get())[]);
extnamestr.get()).index(&FullRange));
return SmallVector::zero();
}
};
@ -913,7 +914,7 @@ fn expand_pat(p: P<ast::Pat>, fld: &mut MacroExpander) -> P<ast::Pat> {
None => {
fld.cx.span_err(pth.span,
format!("macro undefined: '{}!'",
extnamestr)[]);
extnamestr).index(&FullRange));
// let compilation continue
return DummyResult::raw_pat(span);
}
@ -930,11 +931,11 @@ fn expand_pat(p: P<ast::Pat>, fld: &mut MacroExpander) -> P<ast::Pat> {
});
let fm = fresh_mark();
let marked_before = mark_tts(tts[], fm);
let marked_before = mark_tts(tts.index(&FullRange), fm);
let mac_span = fld.cx.original_span();
let expanded = match expander.expand(fld.cx,
mac_span,
marked_before[]).make_pat() {
marked_before.index(&FullRange)).make_pat() {
Some(e) => e,
None => {
fld.cx.span_err(
@ -942,7 +943,7 @@ fn expand_pat(p: P<ast::Pat>, fld: &mut MacroExpander) -> P<ast::Pat> {
format!(
"non-pattern macro in pattern position: {}",
extnamestr.get()
)[]
).index(&FullRange)
);
return DummyResult::raw_pat(span);
}
@ -954,7 +955,7 @@ fn expand_pat(p: P<ast::Pat>, fld: &mut MacroExpander) -> P<ast::Pat> {
_ => {
fld.cx.span_err(span,
format!("{}! is not legal in pattern position",
extnamestr.get())[]);
extnamestr.get()).index(&FullRange));
return DummyResult::raw_pat(span);
}
}
@ -1231,7 +1232,7 @@ impl Folder for Marker {
node: match node {
MacInvocTT(path, tts, ctxt) => {
MacInvocTT(self.fold_path(path),
self.fold_tts(tts[]),
self.fold_tts(tts.index(&FullRange)),
mtwt::apply_mark(self.mark, ctxt))
}
},
@ -1712,7 +1713,7 @@ foo_module!();
let string = ident.get();
"xx" == string
}).collect();
let cxbinds: &[&ast::Ident] = cxbinds[];
let cxbinds: &[&ast::Ident] = cxbinds.index(&FullRange);
let cxbind = match cxbinds {
[b] => b,
_ => panic!("expected just one binding for ext_cx")

View file

@ -113,7 +113,7 @@ fn parse_args(ecx: &mut ExtCtxt, sp: Span, tts: &[ast::TokenTree])
_ => {
ecx.span_err(p.span,
format!("expected ident for named argument, found `{}`",
p.this_token_to_string())[]);
p.this_token_to_string()).index(&FullRange));
return None;
}
};
@ -126,7 +126,7 @@ fn parse_args(ecx: &mut ExtCtxt, sp: Span, tts: &[ast::TokenTree])
Some(prev) => {
ecx.span_err(e.span,
format!("duplicate argument named `{}`",
name)[]);
name).index(&FullRange));
ecx.parse_sess.span_diagnostic.span_note(prev.span, "previously here");
continue
}
@ -217,7 +217,7 @@ impl<'a, 'b> Context<'a, 'b> {
let msg = format!("invalid reference to argument `{}` ({})",
arg, self.describe_num_args());
self.ecx.span_err(self.fmtsp, msg[]);
self.ecx.span_err(self.fmtsp, msg.index(&FullRange));
return;
}
{
@ -237,7 +237,7 @@ impl<'a, 'b> Context<'a, 'b> {
Some(e) => e.span,
None => {
let msg = format!("there is no argument named `{}`", name);
self.ecx.span_err(self.fmtsp, msg[]);
self.ecx.span_err(self.fmtsp, msg.index(&FullRange));
return;
}
};
@ -280,19 +280,19 @@ impl<'a, 'b> Context<'a, 'b> {
format!("argument redeclared with type `{}` when \
it was previously `{}`",
*ty,
*cur)[]);
*cur).index(&FullRange));
}
(&Known(ref cur), _) => {
self.ecx.span_err(sp,
format!("argument used to format with `{}` was \
attempted to not be used for formatting",
*cur)[]);
*cur).index(&FullRange));
}
(_, &Known(ref ty)) => {
self.ecx.span_err(sp,
format!("argument previously used as a format \
argument attempted to be used as `{}`",
*ty)[]);
*ty).index(&FullRange));
}
(_, _) => {
self.ecx.span_err(sp, "argument declared with multiple formats");
@ -357,7 +357,7 @@ impl<'a, 'b> Context<'a, 'b> {
/// Translate the accumulated string literals to a literal expression
fn trans_literal_string(&mut self) -> P<ast::Expr> {
let sp = self.fmtsp;
let s = token::intern_and_get_ident(self.literal[]);
let s = token::intern_and_get_ident(self.literal.index(&FullRange));
self.literal.clear();
self.ecx.expr_str(sp, s)
}
@ -509,7 +509,7 @@ impl<'a, 'b> Context<'a, 'b> {
None => continue // error already generated
};
let name = self.ecx.ident_of(format!("__arg{}", i)[]);
let name = self.ecx.ident_of(format!("__arg{}", i).index(&FullRange));
pats.push(self.ecx.pat_ident(e.span, name));
locals.push(Context::format_arg(self.ecx, e.span, arg_ty,
self.ecx.expr_ident(e.span, name)));
@ -526,7 +526,7 @@ impl<'a, 'b> Context<'a, 'b> {
};
let lname = self.ecx.ident_of(format!("__arg{}",
*name)[]);
*name).index(&FullRange));
pats.push(self.ecx.pat_ident(e.span, lname));
names[self.name_positions[*name]] =
Some(Context::format_arg(self.ecx, e.span, arg_ty,
@ -606,7 +606,7 @@ impl<'a, 'b> Context<'a, 'b> {
-> P<ast::Expr> {
let trait_ = match *ty {
Known(ref tyname) => {
match tyname[] {
match tyname.index(&FullRange) {
"" => "Show",
"?" => "Show",
"e" => "LowerExp",
@ -619,7 +619,7 @@ impl<'a, 'b> Context<'a, 'b> {
_ => {
ecx.span_err(sp,
format!("unknown format trait `{}`",
*tyname)[]);
*tyname).index(&FullRange));
"Dummy"
}
}
@ -710,7 +710,7 @@ pub fn expand_preparsed_format_args(ecx: &mut ExtCtxt, sp: Span,
}
if !parser.errors.is_empty() {
cx.ecx.span_err(cx.fmtsp, format!("invalid format string: {}",
parser.errors.remove(0))[]);
parser.errors.remove(0)).index(&FullRange));
return DummyResult::raw_expr(sp);
}
if !cx.literal.is_empty() {

View file

@ -223,7 +223,7 @@ pub fn marksof(ctxt: SyntaxContext, stopname: Name) -> Vec<Mrk> {
}
// the internal function for computing marks
// it's not clear to me whether it's better to use a [] mutable
// it's not clear to me whether it's better to use a .index(&FullRange) mutable
// vector or a cons-list for this.
fn marksof_internal(ctxt: SyntaxContext,
stopname: Name,

View file

@ -473,7 +473,7 @@ pub fn expand_quote_stmt(cx: &mut ExtCtxt,
}
fn ids_ext(strs: Vec<String> ) -> Vec<ast::Ident> {
strs.iter().map(|str| str_to_ident((*str)[])).collect()
strs.iter().map(|str| str_to_ident((*str).index(&FullRange))).collect()
}
fn id_ext(str: &str) -> ast::Ident {
@ -675,7 +675,7 @@ fn mk_tt(cx: &ExtCtxt, tt: &ast::TokenTree) -> Vec<P<ast::Stmt>> {
for i in range(0, tt.len()) {
seq.push(tt.get_tt(i));
}
mk_tts(cx, seq[])
mk_tts(cx, seq.index(&FullRange))
}
ast::TtToken(sp, ref tok) => {
let e_sp = cx.expr_ident(sp, id_ext("_sp"));
@ -764,7 +764,7 @@ fn expand_tts(cx: &ExtCtxt, sp: Span, tts: &[ast::TokenTree])
let stmt_let_tt = cx.stmt_let(sp, true, id_ext("tt"), cx.expr_vec_ng(sp));
let mut vector = vec!(stmt_let_sp, stmt_let_tt);
vector.extend(mk_tts(cx, tts[]).into_iter());
vector.extend(mk_tts(cx, tts.index(&FullRange)).into_iter());
let block = cx.expr_block(
cx.block_all(sp,
Vec::new(),

View file

@ -57,7 +57,7 @@ pub fn expand_file(cx: &mut ExtCtxt, sp: Span, tts: &[ast::TokenTree])
let topmost = cx.original_span_in_file();
let loc = cx.codemap().lookup_char_pos(topmost.lo);
let filename = token::intern_and_get_ident(loc.file.name[]);
let filename = token::intern_and_get_ident(loc.file.name.index(&FullRange));
base::MacExpr::new(cx.expr_str(topmost, filename))
}
@ -65,7 +65,7 @@ pub fn expand_stringify(cx: &mut ExtCtxt, sp: Span, tts: &[ast::TokenTree])
-> Box<base::MacResult+'static> {
let s = pprust::tts_to_string(tts);
base::MacExpr::new(cx.expr_str(sp,
token::intern_and_get_ident(s[])))
token::intern_and_get_ident(s.index(&FullRange))))
}
pub fn expand_mod(cx: &mut ExtCtxt, sp: Span, tts: &[ast::TokenTree])
@ -78,7 +78,7 @@ pub fn expand_mod(cx: &mut ExtCtxt, sp: Span, tts: &[ast::TokenTree])
.connect("::");
base::MacExpr::new(cx.expr_str(
sp,
token::intern_and_get_ident(string[])))
token::intern_and_get_ident(string.index(&FullRange))))
}
/// include! : parse the given file as an expr
@ -137,7 +137,7 @@ pub fn expand_include_str(cx: &mut ExtCtxt, sp: Span, tts: &[ast::TokenTree])
cx.span_err(sp,
format!("couldn't read {}: {}",
file.display(),
e)[]);
e).index(&FullRange));
return DummyResult::expr(sp);
}
Ok(bytes) => bytes,
@ -147,7 +147,7 @@ pub fn expand_include_str(cx: &mut ExtCtxt, sp: Span, tts: &[ast::TokenTree])
// Add this input file to the code map to make it available as
// dependency information
let filename = file.display().to_string();
let interned = token::intern_and_get_ident(src[]);
let interned = token::intern_and_get_ident(src.index(&FullRange));
cx.codemap().new_filemap(filename, src);
base::MacExpr::new(cx.expr_str(sp, interned))
@ -155,7 +155,7 @@ pub fn expand_include_str(cx: &mut ExtCtxt, sp: Span, tts: &[ast::TokenTree])
Err(_) => {
cx.span_err(sp,
format!("{} wasn't a utf-8 file",
file.display())[]);
file.display()).index(&FullRange));
return DummyResult::expr(sp);
}
}
@ -177,7 +177,7 @@ pub fn expand_include_bytes(cx: &mut ExtCtxt, sp: Span, tts: &[ast::TokenTree])
match File::open(&file).read_to_end() {
Err(e) => {
cx.span_err(sp,
format!("couldn't read {}: {}", file.display(), e)[]);
format!("couldn't read {}: {}", file.display(), e).index(&FullRange));
return DummyResult::expr(sp);
}
Ok(bytes) => {

View file

@ -153,7 +153,7 @@ pub fn count_names(ms: &[TokenTree]) -> uint {
seq.num_captures
}
&TtDelimited(_, ref delim) => {
count_names(delim.tts[])
count_names(delim.tts.index(&FullRange))
}
&TtToken(_, MatchNt(..)) => {
1
@ -165,7 +165,7 @@ pub fn count_names(ms: &[TokenTree]) -> uint {
pub fn initial_matcher_pos(ms: Rc<Vec<TokenTree>>, sep: Option<Token>, lo: BytePos)
-> Box<MatcherPos> {
let match_idx_hi = count_names(ms[]);
let match_idx_hi = count_names(ms.index(&FullRange));
let matches: Vec<_> = range(0, match_idx_hi).map(|_| Vec::new()).collect();
box MatcherPos {
stack: vec![],
@ -229,7 +229,7 @@ pub fn nameize(p_s: &ParseSess, ms: &[TokenTree], res: &[Rc<NamedMatch>])
p_s.span_diagnostic
.span_fatal(sp,
format!("duplicated bind name: {}",
string.get())[])
string.get()).index(&FullRange))
}
}
}
@ -254,13 +254,13 @@ pub fn parse_or_else(sess: &ParseSess,
rdr: TtReader,
ms: Vec<TokenTree> )
-> HashMap<Ident, Rc<NamedMatch>> {
match parse(sess, cfg, rdr, ms[]) {
match parse(sess, cfg, rdr, ms.index(&FullRange)) {
Success(m) => m,
Failure(sp, str) => {
sess.span_diagnostic.span_fatal(sp, str[])
sess.span_diagnostic.span_fatal(sp, str.index(&FullRange))
}
Error(sp, str) => {
sess.span_diagnostic.span_fatal(sp, str[])
sess.span_diagnostic.span_fatal(sp, str.index(&FullRange))
}
}
}
@ -341,7 +341,7 @@ pub fn parse(sess: &ParseSess,
// Only touch the binders we have actually bound
for idx in range(ei.match_lo, ei.match_hi) {
let sub = (ei.matches[idx]).clone();
new_pos.matches[idx]
(&mut new_pos.matches[idx])
.push(Rc::new(MatchedSeq(sub, mk_sp(ei.sp_lo,
sp.hi))));
}
@ -386,7 +386,7 @@ pub fn parse(sess: &ParseSess,
new_ei.idx += 1u;
//we specifically matched zero repeats.
for idx in range(ei.match_cur, ei.match_cur + seq.num_captures) {
new_ei.matches[idx].push(Rc::new(MatchedSeq(Vec::new(), sp)));
(&mut new_ei.matches[idx]).push(Rc::new(MatchedSeq(Vec::new(), sp)));
}
cur_eis.push(new_ei);
@ -444,10 +444,10 @@ pub fn parse(sess: &ParseSess,
if token_name_eq(&tok, &token::Eof) {
if eof_eis.len() == 1u {
let mut v = Vec::new();
for dv in eof_eis[0].matches.iter_mut() {
for dv in (&mut eof_eis[0]).matches.iter_mut() {
v.push(dv.pop().unwrap());
}
return Success(nameize(sess, ms, v[]));
return Success(nameize(sess, ms, v.index(&FullRange)));
} else if eof_eis.len() > 1u {
return Error(sp, "ambiguity: multiple successful parses".to_string());
} else {
@ -486,7 +486,7 @@ pub fn parse(sess: &ParseSess,
TtToken(_, MatchNt(_, name, _, _)) => {
let name_string = token::get_ident(name);
let match_cur = ei.match_cur;
ei.matches[match_cur].push(Rc::new(MatchedNonterminal(
(&mut ei.matches[match_cur]).push(Rc::new(MatchedNonterminal(
parse_nt(&mut rust_parser, name_string.get()))));
ei.idx += 1u;
ei.match_cur += 1;
@ -522,7 +522,7 @@ pub fn parse_nt(p: &mut Parser, name: &str) -> Nonterminal {
_ => {
let token_str = pprust::token_to_string(&p.token);
p.fatal((format!("expected ident, found {}",
token_str[]))[])
token_str.index(&FullRange))).index(&FullRange))
}
},
"path" => {
@ -536,7 +536,7 @@ pub fn parse_nt(p: &mut Parser, name: &str) -> Nonterminal {
res
}
_ => {
p.fatal(format!("unsupported builtin nonterminal parser: {}", name)[])
p.fatal(format!("unsupported builtin nonterminal parser: {}", name).index(&FullRange))
}
}
}

View file

@ -52,7 +52,7 @@ impl<'a> ParserAnyMacro<'a> {
following",
token_str);
let span = parser.span;
parser.span_err(span, msg[]);
parser.span_err(span, msg.index(&FullRange));
}
}
}
@ -126,8 +126,8 @@ impl TTMacroExpander for MacroRulesMacroExpander {
self.name,
self.imported_from,
arg,
self.lhses[],
self.rhses[])
self.lhses.index(&FullRange),
self.rhses.index(&FullRange))
}
}
@ -154,7 +154,7 @@ fn generic_extension<'cx>(cx: &'cx ExtCtxt,
match **lhs {
MatchedNonterminal(NtTT(ref lhs_tt)) => {
let lhs_tt = match **lhs_tt {
TtDelimited(_, ref delim) => delim.tts[],
TtDelimited(_, ref delim) => delim.tts.index(&FullRange),
_ => cx.span_fatal(sp, "malformed macro lhs")
};
// `None` is because we're not interpolating
@ -194,13 +194,13 @@ fn generic_extension<'cx>(cx: &'cx ExtCtxt,
best_fail_spot = sp;
best_fail_msg = (*msg).clone();
},
Error(sp, ref msg) => cx.span_fatal(sp, msg[])
Error(sp, ref msg) => cx.span_fatal(sp, msg.index(&FullRange))
}
}
_ => cx.bug("non-matcher found in parsed lhses")
}
}
cx.span_fatal(best_fail_spot, best_fail_msg[]);
cx.span_fatal(best_fail_spot, best_fail_msg.index(&FullRange));
}
// Note that macro-by-example's input is also matched against a token tree:

View file

@ -240,7 +240,7 @@ pub fn tt_next_token(r: &mut TtReader) -> TokenAndSpan {
}
LisContradiction(ref msg) => {
// FIXME #2887 blame macro invoker instead
r.sp_diag.span_fatal(sp.clone(), msg[]);
r.sp_diag.span_fatal(sp.clone(), msg.index(&FullRange));
}
LisConstraint(len, _) => {
if len == 0 {
@ -297,7 +297,7 @@ pub fn tt_next_token(r: &mut TtReader) -> TokenAndSpan {
r.sp_diag.span_fatal(
r.cur_span, /* blame the macro writer */
format!("variable '{}' is still repeating at this depth",
token::get_ident(ident))[]);
token::get_ident(ident)).index(&FullRange));
}
}
}