aboutsummaryrefslogtreecommitdiff
path: root/src/comp
diff options
context:
space:
mode:
authorMarijn Haverbeke <[email protected]>2011-05-12 17:20:07 +0200
committerMarijn Haverbeke <[email protected]>2011-05-12 21:30:35 +0200
commit6510f1ce7cb026ceb98689583c13f18c1e7c2c12 (patch)
treedaee2983e121887082fbba929919d2a2dccef36c /src/comp
parentRemove some truly wrong logic in parse_constrs (diff)
downloadrust-6510f1ce7cb026ceb98689583c13f18c1e7c2c12.tar.xz
rust-6510f1ce7cb026ceb98689583c13f18c1e7c2c12.zip
Change module dereference syntax from . to ::
This will need to be a snapshot.
Diffstat (limited to 'src/comp')
-rw-r--r--src/comp/front/creader.rs2
-rw-r--r--src/comp/front/lexer.rs12
-rw-r--r--src/comp/front/parser.rs72
-rw-r--r--src/comp/front/token.rs2
-rw-r--r--src/comp/middle/metadata.rs2
-rw-r--r--src/comp/middle/resolve.rs54
-rw-r--r--src/comp/middle/trans.rs2
-rw-r--r--src/comp/middle/ty.rs2
-rw-r--r--src/comp/pretty/pprust.rs4
9 files changed, 41 insertions, 111 deletions
diff --git a/src/comp/front/creader.rs b/src/comp/front/creader.rs
index dd913b53..5ae56417 100644
--- a/src/comp/front/creader.rs
+++ b/src/comp/front/creader.rs
@@ -312,7 +312,7 @@ fn resolve_path(vec[ast.ident] path, vec[u8] data) -> resolve_result {
fn eq_item(vec[u8] data, str s) -> bool {
ret Str.eq(Str.unsafe_from_bytes(data), s);
}
- auto s = Str.connect(path, ".");
+ auto s = Str.connect(path, "::");
auto md = EBML.new_doc(data);
auto paths = EBML.get_doc(md, metadata.tag_paths);
auto eqer = bind eq_item(_, s);
diff --git a/src/comp/front/lexer.rs b/src/comp/front/lexer.rs
index 32f21531..adcdaa16 100644
--- a/src/comp/front/lexer.rs
+++ b/src/comp/front/lexer.rs
@@ -634,7 +634,6 @@ fn next_token(reader rdr) -> token.token {
alt (c) {
// One-byte tokens.
- case (':') { rdr.bump(); ret token.COLON; }
case ('?') { rdr.bump(); ret token.QUES; }
case (';') { rdr.bump(); ret token.SEMI; }
case (',') { rdr.bump(); ret token.COMMA; }
@@ -648,7 +647,16 @@ fn next_token(reader rdr) -> token.token {
case ('@') { rdr.bump(); ret token.AT; }
case ('#') { rdr.bump(); ret token.POUND; }
case ('~') { rdr.bump(); ret token.TILDE; }
-
+ case (':') {
+ rdr.bump();
+ if (rdr.curr() == ':') {
+ rdr.bump();
+ ret token.MOD_SEP;
+ }
+ else {
+ ret token.COLON;
+ };
+ }
// Multi-byte tokens.
case ('=') {
diff --git a/src/comp/front/parser.rs b/src/comp/front/parser.rs
index 2295e478..a14bcf3a 100644
--- a/src/comp/front/parser.rs
+++ b/src/comp/front/parser.rs
@@ -322,7 +322,7 @@ fn parse_constr_arg(parser p) -> @ast.constr_arg {
fn parse_ty_constr(parser p) -> @ast.constr {
auto lo = p.get_lo_pos();
- auto path = parse_path(p, GREEDY);
+ auto path = parse_path(p);
auto pf = parse_constr_arg;
auto args = parse_seq[@ast.constr_arg](token.LPAREN,
token.RPAREN,
@@ -472,7 +472,7 @@ fn parse_ty(parser p) -> @ast.ty {
}
case (token.IDENT(_)) {
- auto path = parse_path(p, GREEDY);
+ auto path = parse_path(p);
t = ast.ty_path(path, p.get_ann());
hi = path.span.hi;
}
@@ -603,11 +603,6 @@ fn is_ident(token.token t) -> bool {
ret false;
}
-tag greed {
- GREEDY;
- MINIMAL;
-}
-
fn parse_ty_args(parser p, uint hi) ->
util.common.spanned[vec[@ast.ty]] {
@@ -623,33 +618,23 @@ fn parse_ty_args(parser p, uint hi) ->
ret spanned(hi, hi, v);
}
-fn parse_path(parser p, greed g) -> ast.path {
+fn parse_path(parser p) -> ast.path {
auto lo = p.get_lo_pos();
auto hi = lo;
let vec[ast.ident] ids = vec();
- let bool more = true;
- while (more) {
+ while (true) {
alt (p.peek()) {
case (token.IDENT(?i)) {
hi = p.get_hi_pos();
ids += vec(p.get_str(i));
p.bump();
- if (p.peek() == token.DOT) {
- if (g == GREEDY) {
- p.bump();
- assert (is_ident(p.peek()));
- } else {
- more = false;
- }
- } else {
- more = false;
- }
- }
- case (_) {
- more = false;
+ if (p.peek() == token.MOD_SEP) {
+ p.bump();
+ } else { break; }
}
+ case (_) { break; }
}
}
@@ -690,7 +675,7 @@ fn parse_bottom_expr(parser p) -> @ast.expr {
alt (p.peek()) {
case (token.IDENT(_)) {
- auto pth = parse_path(p, MINIMAL);
+ auto pth = parse_path(p);
hi = pth.span.hi;
ex = ast.expr_path(pth, p.get_ann());
}
@@ -804,7 +789,7 @@ fn parse_bottom_expr(parser p) -> @ast.expr {
case (token.POUND) {
p.bump();
- auto pth = parse_path(p, GREEDY);
+ auto pth = parse_path(p);
auto pf = parse_expr;
auto es = parse_seq[@ast.expr](token.LPAREN,
token.RPAREN,
@@ -980,31 +965,6 @@ fn expand_syntax_ext(parser p, ast.span sp,
}
}
-fn extend_expr_by_ident(parser p, uint lo, uint hi,
- @ast.expr e, ast.ident i) -> @ast.expr {
- auto e_ = e.node;
- alt (e.node) {
- case (ast.expr_path(?pth, ?ann)) {
- if (Vec.len[@ast.ty](pth.node.types) == 0u) {
- auto idents_ = pth.node.idents;
- idents_ += vec(i);
- auto tys = parse_ty_args(p, hi);
- auto pth_ = spanned(pth.span.lo, tys.span.hi,
- rec(idents=idents_,
- types=tys.node));
- e_ = ast.expr_path(pth_, ann);
- ret @spanned(pth_.span.lo, pth_.span.hi, e_);
- } else {
- e_ = ast.expr_field(e, i, ann);
- }
- }
- case (_) {
- e_ = ast.expr_field(e, i, p.get_ann());
- }
- }
- ret @spanned(lo, hi, e_);
-}
-
fn parse_self_method(parser p) -> @ast.expr {
auto sp = p.get_span();
let ast.ident f_name = parse_ident(p);
@@ -1042,7 +1002,9 @@ fn parse_dot_or_call_expr(parser p) -> @ast.expr {
case (token.IDENT(?i)) {
hi = p.get_hi_pos();
p.bump();
- e = extend_expr_by_ident(p, lo, hi, e, p.get_str(i));
+ auto e_ = ast.expr_field(e, p.get_str(i),
+ p.get_ann());
+ e = @spanned(lo, hi, e_);
}
case (token.LPAREN) {
@@ -1404,7 +1366,7 @@ fn parse_spawn_expr(parser p) -> @ast.expr {
expect(p, token.SPAWN);
// FIXME: Parse domain and name
-
+ // FIXME: why no full expr?
auto fn_expr = parse_bottom_expr(p);
auto pf = parse_expr;
auto es = parse_seq[@ast.expr](token.LPAREN,
@@ -1509,7 +1471,7 @@ fn parse_pat(parser p) -> @ast.pat {
}
}
case (token.IDENT(_)) {
- auto tag_path = parse_path(p, GREEDY);
+ auto tag_path = parse_path(p);
hi = tag_path.span.hi;
let vec[@ast.pat] args;
@@ -2249,7 +2211,7 @@ fn parse_rest_import_name(parser p, ast.ident first,
auto lo = p.get_lo_pos();
let vec[ast.ident] identifiers = vec(first);
while (p.peek() != token.SEMI) {
- expect(p, token.DOT);
+ expect(p, token.MOD_SEP);
auto i = parse_ident(p);
identifiers += vec(i);
}
@@ -2377,7 +2339,7 @@ fn parse_crate_directive(parser p) -> ast.crate_directive
alt (p.peek()) {
case (token.AUTH) {
p.bump();
- auto n = parse_path(p, GREEDY);
+ auto n = parse_path(p);
expect(p, token.EQ);
auto a = parse_auth(p);
auto hi = p.get_hi_pos();
diff --git a/src/comp/front/token.rs b/src/comp/front/token.rs
index 4c2891f3..43db0b16 100644
--- a/src/comp/front/token.rs
+++ b/src/comp/front/token.rs
@@ -47,6 +47,7 @@ tag token {
COMMA;
SEMI;
COLON;
+ MOD_SEP;
QUES;
RARROW;
SEND;
@@ -218,6 +219,7 @@ fn to_str(lexer.reader r, token t) -> str {
case (COMMA) { ret ","; }
case (SEMI) { ret ";"; }
case (COLON) { ret ":"; }
+ case (MOD_SEP) { ret "::"; }
case (QUES) { ret "?"; }
case (RARROW) { ret "->"; }
case (SEND) { ret "<|"; }
diff --git a/src/comp/middle/metadata.rs b/src/comp/middle/metadata.rs
index eedff3bf..ee628643 100644
--- a/src/comp/middle/metadata.rs
+++ b/src/comp/middle/metadata.rs
@@ -301,7 +301,7 @@ fn add_to_index(&EBML.writer ebml_w,
&mutable vec[tup(str, uint)] index,
&str name) {
auto full_path = path + vec(name);
- index += vec(tup(Str.connect(full_path, "."), ebml_w.writer.tell()));
+ index += vec(tup(Str.connect(full_path, "::"), ebml_w.writer.tell()));
}
fn encode_native_module_item_paths(&EBML.writer ebml_w,
diff --git a/src/comp/middle/resolve.rs b/src/comp/middle/resolve.rs
index 1305ad5f..7b7d6369 100644
--- a/src/comp/middle/resolve.rs
+++ b/src/comp/middle/resolve.rs
@@ -170,6 +170,7 @@ fn resolve_imports(&env e) {
}
}
+// FIXME this should use walk (will need to add walk_arm)
fn resolve_names(&@env e, &ast.crate c) -> @ast.crate {
auto fld = @rec(fold_pat_tag = bind fold_pat_tag(e,_,_,_,_,_),
fold_expr_path = bind fold_expr_path(e,_,_,_,_),
@@ -277,54 +278,11 @@ fn resolve_import(&env e, &@ast.view_item it, &list[scope] sc) {
}
}
-// We received a path expression of the following form:
-//
-// a.b.c.d
-//
-// Somewhere along this path there might be a split from a path-expr
-// to a runtime field-expr. For example:
-//
-// 'a' could be the name of a variable in the local scope
-// and 'b.c.d' could be a field-sequence inside it.
-//
-// Or:
-//
-// 'a.b' could be a module path to a constant record, and 'c.d'
-// could be a field within it.
-//
-// Our job here is to figure out what the prefix of 'a.b.c.d' is that
-// corresponds to a static binding-name (a module or slot, with no type info)
-// and split that off as the 'primary' expr_path, with secondary expr_field
-// expressions tacked on the end.
-
fn fold_expr_path(@env e, &list[scope] sc, &span sp, &ast.path p, &ann a)
-> @ast.expr {
- auto idents = p.node.idents;
- auto n_idents = Vec.len(idents);
- assert (n_idents != 0u);
-
- auto dcur = lookup_in_scope_strict(*e, sc, sp, idents.(0), ns_value);
- auto i = 1u;
- while (i < n_idents) {
- if (!is_module(dcur)) { break; }
- dcur = lookup_in_mod_strict(*e, dcur, sp, idents.(i), ns_value,
- outside);
- i += 1u;
- }
- if (is_module(dcur)) {
- e.sess.span_err(sp, "can't refer to a module as a first-class value");
- }
-
- p = rec(node=rec(idents=Vec.slice(idents, 0u, i) with p.node) with p);
- auto ex = @fold.respan(sp, ast.expr_path(p, a));
- e.def_map.insert(ast.ann_tag(a), dcur);
- // FIXME this duplicates the ann. Is that a problem? How will we deal with
- // splitting this into path and field exprs when we don't fold?
- while (i < n_idents) {
- ex = @fold.respan(sp, ast.expr_field(ex, idents.(i), a));
- i += 1u;
- }
- ret ex;
+ auto df = lookup_path_strict(*e, sc, sp, p.node.idents, ns_value);
+ e.def_map.insert(ast.ann_tag(a), df);
+ ret @fold.respan(sp, ast.expr_path(p, a));
}
@@ -337,7 +295,7 @@ fn fold_pat_tag(@env e, &list[scope] sc, &span sp, &ast.path p,
}
case (_) {
e.sess.span_err(sp, "not a tag variant: " +
- Str.connect(p.node.idents, "."));
+ Str.connect(p.node.idents, "::"));
fail;
}
}
@@ -383,7 +341,7 @@ fn lookup_path_strict(&env e, &list[scope] sc, &span sp, vec[ident] idents,
i += 1u;
}
if (is_module(dcur)) {
- e.sess.span_err(sp, Str.connect(idents, ".") +
+ e.sess.span_err(sp, Str.connect(idents, "::") +
" is a module, not a " + ns_name(ns));
}
ret dcur;
diff --git a/src/comp/middle/trans.rs b/src/comp/middle/trans.rs
index d6e86045..ecdf78cf 100644
--- a/src/comp/middle/trans.rs
+++ b/src/comp/middle/trans.rs
@@ -5439,7 +5439,7 @@ fn load_if_immediate(&@block_ctxt cx, ValueRef v, &ty.t t) -> ValueRef {
fn trans_log(int lvl, &@block_ctxt cx, &@ast.expr e) -> result {
auto lcx = cx.fcx.lcx;
- auto modname = Str.connect(lcx.module_path, ".");
+ auto modname = Str.connect(lcx.module_path, "::");
auto global;
if (lcx.ccx.module_data.contains_key(modname)) {
global = lcx.ccx.module_data.get(modname);
diff --git a/src/comp/middle/ty.rs b/src/comp/middle/ty.rs
index c749b220..8c79c605 100644
--- a/src/comp/middle/ty.rs
+++ b/src/comp/middle/ty.rs
@@ -484,7 +484,7 @@ fn cname(&ctxt cx, &t typ) -> Option.t[str] { ret cx.ts.others.(typ).cname; }
// Stringification
fn path_to_str(&ast.path pth) -> str {
- auto result = Str.connect(pth.node.idents, ".");
+ auto result = Str.connect(pth.node.idents, "::");
if (Vec.len[@ast.ty](pth.node.types) > 0u) {
auto f = pretty.pprust.ty_to_str;
result += "[";
diff --git a/src/comp/pretty/pprust.rs b/src/comp/pretty/pprust.rs
index d380f0b7..071e076c 100644
--- a/src/comp/pretty/pprust.rs
+++ b/src/comp/pretty/pprust.rs
@@ -751,7 +751,7 @@ fn print_path(ps s, ast.path path) {
auto first = true;
for (str id in path.node.idents) {
if (first) {first = false;}
- else {wrd(s.s, ".");}
+ else {wrd(s.s, "::");}
wrd(s.s, id);
}
if (Vec.len[@ast.ty](path.node.types) > 0u) {
@@ -856,7 +856,7 @@ fn print_view_item(ps s, @ast.view_item item) {
auto first = true;
for (str elt in ids) {
if (first) {first = false;}
- else {wrd(s.s, ".");}
+ else {wrd(s.s, ":");}
wrd(s.s, elt);
}
}