1
  2
  3
  4
  5
  6
  7
  8
  9
 10
 11
 12
 13
 14
 15
 16
 17
 18
 19
 20
 21
 22
 23
 24
 25
 26
 27
 28
 29
 30
 31
 32
 33
 34
 35
 36
 37
 38
 39
 40
 41
 42
 43
 44
 45
 46
 47
 48
 49
 50
 51
 52
 53
 54
 55
 56
 57
 58
 59
 60
 61
 62
 63
 64
 65
 66
 67
 68
 69
 70
 71
 72
 73
 74
 75
 76
 77
 78
 79
 80
 81
 82
 83
 84
 85
 86
 87
 88
 89
 90
 91
 92
 93
 94
 95
 96
 97
 98
 99
100
use read_token::ReadToken;
use range::Range;
use std::sync::Arc;

use super::{
    ParseResult,
};
use {
    DebugId,
    MetaData,
    ParseError,
};
use tokenizer::{ read_data, TokenizerState };

/// Stores information about reading until whitespace or any of some character.
#[derive(Clone, Debug, PartialEq)]
pub struct UntilAnyOrWhitespace {
    /// The characters to stop at.
    pub any_characters: Arc<String>,
    /// Whether empty data is accepted or not.
    pub optional: bool,
    /// The property to store read text.
    pub property: Option<Arc<String>>,
    /// A debug id to track down the rule generating an error.
    pub debug_id: DebugId,
}

impl UntilAnyOrWhitespace {
    /// Parses until whitespace or any specified characters.
    pub fn parse(
        &self,
        tokens: &mut Vec<Range<MetaData>>,
        state: &TokenizerState,
        read_token: &ReadToken
    ) -> ParseResult<TokenizerState> {
        let (range, _) = read_token.until_any_or_whitespace(
            &self.any_characters);
        if range.length == 0 && !self.optional {
            Err(range.wrap(ParseError::ExpectedSomething(self.debug_id)))
        } else {
            if let Some(ref property) = self.property {
                let text = read_token.raw_string(range.length);
                Ok((range, read_data(
                    tokens,
                    range.wrap(
                        MetaData::String(property.clone(), Arc::new(text))),
                    state
                ), None))
            } else {
                Ok((range, state.clone(), None))
            }
        }
    }
}

#[cfg(test)]
mod tests {
    use all::*;
    use all::tokenizer::*;
    use meta_rules::UntilAnyOrWhitespace;
    use range::Range;
    use read_token::ReadToken;
    use std::sync::Arc;

    #[test]
    fn required() {
        let text = "fn ()";
        let mut tokenizer = vec![];
        let s = TokenizerState::new();
        let name = UntilAnyOrWhitespace {
            debug_id: 0,
            any_characters: Arc::new("(".into()),
            optional: false,
            property: None
        };
        let res = name.parse(&mut tokenizer, &s,
            &ReadToken::new(&text[3..], 3));
        assert_eq!(res, Err(Range::new(3, 0).wrap(
            ParseError::ExpectedSomething(0))));
    }

    #[test]
    fn successful() {
        let text = "fn foo()";
        let mut tokens = vec![];
        let s = TokenizerState::new();
        let function_name: Arc<String> = Arc::new("function_name".into());
        let name = UntilAnyOrWhitespace {
            debug_id: 0,
            any_characters: Arc::new("(".into()),
            optional: false,
            property: Some(function_name.clone())
        };
        let res = name.parse(&mut tokens, &s, &ReadToken::new(&text[3..], 3));
        assert_eq!(res, Ok((Range::new(3, 3), TokenizerState(1), None)));
        assert_eq!(tokens.len(), 1);
        assert_eq!(&tokens[0].data,
            &MetaData::String(function_name.clone(), Arc::new("foo".into())));
    }
}