[898] | 1 | |
---|
| 2 | # The following YAML grammar is LL(1) and is parsed by a recursive descent |
---|
| 3 | # parser. |
---|
| 4 | # |
---|
| 5 | # stream ::= STREAM-START implicit_document? explicit_document* STREAM-END |
---|
| 6 | # implicit_document ::= block_node DOCUMENT-END* |
---|
| 7 | # explicit_document ::= DIRECTIVE* DOCUMENT-START block_node? DOCUMENT-END* |
---|
| 8 | # block_node_or_indentless_sequence ::= |
---|
| 9 | # ALIAS |
---|
| 10 | # | properties (block_content | indentless_block_sequence)? |
---|
| 11 | # | block_content |
---|
| 12 | # | indentless_block_sequence |
---|
| 13 | # block_node ::= ALIAS |
---|
| 14 | # | properties block_content? |
---|
| 15 | # | block_content |
---|
| 16 | # flow_node ::= ALIAS |
---|
| 17 | # | properties flow_content? |
---|
| 18 | # | flow_content |
---|
| 19 | # properties ::= TAG ANCHOR? | ANCHOR TAG? |
---|
| 20 | # block_content ::= block_collection | flow_collection | SCALAR |
---|
| 21 | # flow_content ::= flow_collection | SCALAR |
---|
| 22 | # block_collection ::= block_sequence | block_mapping |
---|
| 23 | # flow_collection ::= flow_sequence | flow_mapping |
---|
| 24 | # block_sequence ::= BLOCK-SEQUENCE-START (BLOCK-ENTRY block_node?)* BLOCK-END |
---|
| 25 | # indentless_sequence ::= (BLOCK-ENTRY block_node?)+ |
---|
| 26 | # block_mapping ::= BLOCK-MAPPING_START |
---|
| 27 | # ((KEY block_node_or_indentless_sequence?)? |
---|
| 28 | # (VALUE block_node_or_indentless_sequence?)?)* |
---|
| 29 | # BLOCK-END |
---|
| 30 | # flow_sequence ::= FLOW-SEQUENCE-START |
---|
| 31 | # (flow_sequence_entry FLOW-ENTRY)* |
---|
| 32 | # flow_sequence_entry? |
---|
| 33 | # FLOW-SEQUENCE-END |
---|
| 34 | # flow_sequence_entry ::= flow_node | KEY flow_node? (VALUE flow_node?)? |
---|
| 35 | # flow_mapping ::= FLOW-MAPPING-START |
---|
| 36 | # (flow_mapping_entry FLOW-ENTRY)* |
---|
| 37 | # flow_mapping_entry? |
---|
| 38 | # FLOW-MAPPING-END |
---|
| 39 | # flow_mapping_entry ::= flow_node | KEY flow_node? (VALUE flow_node?)? |
---|
| 40 | # |
---|
| 41 | # FIRST sets: |
---|
| 42 | # |
---|
| 43 | # stream: { STREAM-START } |
---|
| 44 | # explicit_document: { DIRECTIVE DOCUMENT-START } |
---|
| 45 | # implicit_document: FIRST(block_node) |
---|
| 46 | # block_node: { ALIAS TAG ANCHOR SCALAR BLOCK-SEQUENCE-START BLOCK-MAPPING-START FLOW-SEQUENCE-START FLOW-MAPPING-START } |
---|
| 47 | # flow_node: { ALIAS ANCHOR TAG SCALAR FLOW-SEQUENCE-START FLOW-MAPPING-START } |
---|
| 48 | # block_content: { BLOCK-SEQUENCE-START BLOCK-MAPPING-START FLOW-SEQUENCE-START FLOW-MAPPING-START SCALAR } |
---|
| 49 | # flow_content: { FLOW-SEQUENCE-START FLOW-MAPPING-START SCALAR } |
---|
| 50 | # block_collection: { BLOCK-SEQUENCE-START BLOCK-MAPPING-START } |
---|
| 51 | # flow_collection: { FLOW-SEQUENCE-START FLOW-MAPPING-START } |
---|
| 52 | # block_sequence: { BLOCK-SEQUENCE-START } |
---|
| 53 | # block_mapping: { BLOCK-MAPPING-START } |
---|
| 54 | # block_node_or_indentless_sequence: { ALIAS ANCHOR TAG SCALAR BLOCK-SEQUENCE-START BLOCK-MAPPING-START FLOW-SEQUENCE-START FLOW-MAPPING-START BLOCK-ENTRY } |
---|
| 55 | # indentless_sequence: { ENTRY } |
---|
| 56 | # flow_collection: { FLOW-SEQUENCE-START FLOW-MAPPING-START } |
---|
| 57 | # flow_sequence: { FLOW-SEQUENCE-START } |
---|
| 58 | # flow_mapping: { FLOW-MAPPING-START } |
---|
| 59 | # flow_sequence_entry: { ALIAS ANCHOR TAG SCALAR FLOW-SEQUENCE-START FLOW-MAPPING-START KEY } |
---|
| 60 | # flow_mapping_entry: { ALIAS ANCHOR TAG SCALAR FLOW-SEQUENCE-START FLOW-MAPPING-START KEY } |
---|
| 61 | |
---|
| 62 | __all__ = ['Parser', 'ParserError'] |
---|
| 63 | |
---|
| 64 | from error import MarkedYAMLError |
---|
| 65 | from tokens import * |
---|
| 66 | from events import * |
---|
| 67 | from scanner import * |
---|
| 68 | |
---|
| 69 | class ParserError(MarkedYAMLError): |
---|
| 70 | pass |
---|
| 71 | |
---|
| 72 | class Parser(object): |
---|
| 73 | # Since writing a recursive-descendant parser is a straightforward task, we |
---|
| 74 | # do not give many comments here. |
---|
| 75 | # Note that we use Python generators. If you rewrite the parser in another |
---|
| 76 | # language, you may replace all 'yield'-s with event handler calls. |
---|
| 77 | |
---|
| 78 | DEFAULT_TAGS = { |
---|
| 79 | u'!': u'!', |
---|
| 80 | u'!!': u'tag:yaml.org,2002:', |
---|
| 81 | } |
---|
| 82 | |
---|
| 83 | def __init__(self): |
---|
| 84 | self.current_event = None |
---|
| 85 | self.yaml_version = None |
---|
| 86 | self.tag_handles = {} |
---|
| 87 | self.states = [] |
---|
| 88 | self.marks = [] |
---|
| 89 | self.state = self.parse_stream_start |
---|
| 90 | |
---|
| 91 | def check_event(self, *choices): |
---|
| 92 | # Check the type of the next event. |
---|
| 93 | if self.current_event is None: |
---|
| 94 | if self.state: |
---|
| 95 | self.current_event = self.state() |
---|
| 96 | if self.current_event is not None: |
---|
| 97 | if not choices: |
---|
| 98 | return True |
---|
| 99 | for choice in choices: |
---|
| 100 | if isinstance(self.current_event, choice): |
---|
| 101 | return True |
---|
| 102 | return False |
---|
| 103 | |
---|
| 104 | def peek_event(self): |
---|
| 105 | # Get the next event. |
---|
| 106 | if self.current_event is None: |
---|
| 107 | if self.state: |
---|
| 108 | self.current_event = self.state() |
---|
| 109 | return self.current_event |
---|
| 110 | |
---|
| 111 | def get_event(self): |
---|
| 112 | # Get the next event and proceed further. |
---|
| 113 | if self.current_event is None: |
---|
| 114 | if self.state: |
---|
| 115 | self.current_event = self.state() |
---|
| 116 | value = self.current_event |
---|
| 117 | self.current_event = None |
---|
| 118 | return value |
---|
| 119 | |
---|
| 120 | # stream ::= STREAM-START implicit_document? explicit_document* STREAM-END |
---|
| 121 | # implicit_document ::= block_node DOCUMENT-END* |
---|
| 122 | # explicit_document ::= DIRECTIVE* DOCUMENT-START block_node? DOCUMENT-END* |
---|
| 123 | |
---|
| 124 | def parse_stream_start(self): |
---|
| 125 | |
---|
| 126 | # Parse the stream start. |
---|
| 127 | token = self.get_token() |
---|
| 128 | event = StreamStartEvent(token.start_mark, token.end_mark, |
---|
| 129 | encoding=token.encoding) |
---|
| 130 | |
---|
| 131 | # Prepare the next state. |
---|
| 132 | self.state = self.parse_implicit_document_start |
---|
| 133 | |
---|
| 134 | return event |
---|
| 135 | |
---|
| 136 | def parse_implicit_document_start(self): |
---|
| 137 | |
---|
| 138 | # Parse an implicit document. |
---|
| 139 | if not self.check_token(DirectiveToken, DocumentStartToken, |
---|
| 140 | StreamEndToken): |
---|
| 141 | self.tag_handles = self.DEFAULT_TAGS |
---|
| 142 | token = self.peek_token() |
---|
| 143 | start_mark = end_mark = token.start_mark |
---|
| 144 | event = DocumentStartEvent(start_mark, end_mark, |
---|
| 145 | explicit=False) |
---|
| 146 | |
---|
| 147 | # Prepare the next state. |
---|
| 148 | self.states.append(self.parse_document_end) |
---|
| 149 | self.state = self.parse_block_node |
---|
| 150 | |
---|
| 151 | return event |
---|
| 152 | |
---|
| 153 | else: |
---|
| 154 | return self.parse_document_start() |
---|
| 155 | |
---|
| 156 | def parse_document_start(self): |
---|
| 157 | |
---|
| 158 | # Parse any extra document end indicators. |
---|
| 159 | while self.check_token(DocumentEndToken): |
---|
| 160 | self.get_token() |
---|
| 161 | |
---|
| 162 | # Parse an explicit document. |
---|
| 163 | if not self.check_token(StreamEndToken): |
---|
| 164 | token = self.peek_token() |
---|
| 165 | start_mark = token.start_mark |
---|
| 166 | version, tags = self.process_directives() |
---|
| 167 | if not self.check_token(DocumentStartToken): |
---|
| 168 | raise ParserError(None, None, |
---|
| 169 | "expected '<document start>', but found %r" |
---|
| 170 | % self.peek_token().id, |
---|
| 171 | self.peek_token().start_mark) |
---|
| 172 | token = self.get_token() |
---|
| 173 | end_mark = token.end_mark |
---|
| 174 | event = DocumentStartEvent(start_mark, end_mark, |
---|
| 175 | explicit=True, version=version, tags=tags) |
---|
| 176 | self.states.append(self.parse_document_end) |
---|
| 177 | self.state = self.parse_document_content |
---|
| 178 | else: |
---|
| 179 | # Parse the end of the stream. |
---|
| 180 | token = self.get_token() |
---|
| 181 | event = StreamEndEvent(token.start_mark, token.end_mark) |
---|
| 182 | assert not self.states |
---|
| 183 | assert not self.marks |
---|
| 184 | self.state = None |
---|
| 185 | return event |
---|
| 186 | |
---|
| 187 | def parse_document_end(self): |
---|
| 188 | |
---|
| 189 | # Parse the document end. |
---|
| 190 | token = self.peek_token() |
---|
| 191 | start_mark = end_mark = token.start_mark |
---|
| 192 | explicit = False |
---|
| 193 | if self.check_token(DocumentEndToken): |
---|
| 194 | token = self.get_token() |
---|
| 195 | end_mark = token.end_mark |
---|
| 196 | explicit = True |
---|
| 197 | event = DocumentEndEvent(start_mark, end_mark, |
---|
| 198 | explicit=explicit) |
---|
| 199 | |
---|
| 200 | # Prepare the next state. |
---|
| 201 | self.state = self.parse_document_start |
---|
| 202 | |
---|
| 203 | return event |
---|
| 204 | |
---|
| 205 | def parse_document_content(self): |
---|
| 206 | if self.check_token(DirectiveToken, |
---|
| 207 | DocumentStartToken, DocumentEndToken, StreamEndToken): |
---|
| 208 | event = self.process_empty_scalar(self.peek_token().start_mark) |
---|
| 209 | self.state = self.states.pop() |
---|
| 210 | return event |
---|
| 211 | else: |
---|
| 212 | return self.parse_block_node() |
---|
| 213 | |
---|
| 214 | def process_directives(self): |
---|
| 215 | self.yaml_version = None |
---|
| 216 | self.tag_handles = {} |
---|
| 217 | while self.check_token(DirectiveToken): |
---|
| 218 | token = self.get_token() |
---|
| 219 | if token.name == u'YAML': |
---|
| 220 | if self.yaml_version is not None: |
---|
| 221 | raise ParserError(None, None, |
---|
| 222 | "found duplicate YAML directive", token.start_mark) |
---|
| 223 | major, minor = token.value |
---|
| 224 | if major != 1: |
---|
| 225 | raise ParserError(None, None, |
---|
| 226 | "found incompatible YAML document (version 1.* is required)", |
---|
| 227 | token.start_mark) |
---|
| 228 | self.yaml_version = token.value |
---|
| 229 | elif token.name == u'TAG': |
---|
| 230 | handle, prefix = token.value |
---|
| 231 | if handle in self.tag_handles: |
---|
| 232 | raise ParserError(None, None, |
---|
| 233 | "duplicate tag handle %r" % handle.encode('utf-8'), |
---|
| 234 | token.start_mark) |
---|
| 235 | self.tag_handles[handle] = prefix |
---|
| 236 | if self.tag_handles: |
---|
| 237 | value = self.yaml_version, self.tag_handles.copy() |
---|
| 238 | else: |
---|
| 239 | value = self.yaml_version, None |
---|
| 240 | for key in self.DEFAULT_TAGS: |
---|
| 241 | if key not in self.tag_handles: |
---|
| 242 | self.tag_handles[key] = self.DEFAULT_TAGS[key] |
---|
| 243 | return value |
---|
| 244 | |
---|
| 245 | # block_node_or_indentless_sequence ::= ALIAS |
---|
| 246 | # | properties (block_content | indentless_block_sequence)? |
---|
| 247 | # | block_content |
---|
| 248 | # | indentless_block_sequence |
---|
| 249 | # block_node ::= ALIAS |
---|
| 250 | # | properties block_content? |
---|
| 251 | # | block_content |
---|
| 252 | # flow_node ::= ALIAS |
---|
| 253 | # | properties flow_content? |
---|
| 254 | # | flow_content |
---|
| 255 | # properties ::= TAG ANCHOR? | ANCHOR TAG? |
---|
| 256 | # block_content ::= block_collection | flow_collection | SCALAR |
---|
| 257 | # flow_content ::= flow_collection | SCALAR |
---|
| 258 | # block_collection ::= block_sequence | block_mapping |
---|
| 259 | # flow_collection ::= flow_sequence | flow_mapping |
---|
| 260 | |
---|
| 261 | def parse_block_node(self): |
---|
| 262 | return self.parse_node(block=True) |
---|
| 263 | |
---|
| 264 | def parse_flow_node(self): |
---|
| 265 | return self.parse_node() |
---|
| 266 | |
---|
| 267 | def parse_block_node_or_indentless_sequence(self): |
---|
| 268 | return self.parse_node(block=True, indentless_sequence=True) |
---|
| 269 | |
---|
| 270 | def parse_node(self, block=False, indentless_sequence=False): |
---|
| 271 | if self.check_token(AliasToken): |
---|
| 272 | token = self.get_token() |
---|
| 273 | event = AliasEvent(token.value, token.start_mark, token.end_mark) |
---|
| 274 | self.state = self.states.pop() |
---|
| 275 | else: |
---|
| 276 | anchor = None |
---|
| 277 | tag = None |
---|
| 278 | start_mark = end_mark = tag_mark = None |
---|
| 279 | if self.check_token(AnchorToken): |
---|
| 280 | token = self.get_token() |
---|
| 281 | start_mark = token.start_mark |
---|
| 282 | end_mark = token.end_mark |
---|
| 283 | anchor = token.value |
---|
| 284 | if self.check_token(TagToken): |
---|
| 285 | token = self.get_token() |
---|
| 286 | tag_mark = token.start_mark |
---|
| 287 | end_mark = token.end_mark |
---|
| 288 | tag = token.value |
---|
| 289 | elif self.check_token(TagToken): |
---|
| 290 | token = self.get_token() |
---|
| 291 | start_mark = tag_mark = token.start_mark |
---|
| 292 | end_mark = token.end_mark |
---|
| 293 | tag = token.value |
---|
| 294 | if self.check_token(AnchorToken): |
---|
| 295 | token = self.get_token() |
---|
| 296 | end_mark = token.end_mark |
---|
| 297 | anchor = token.value |
---|
| 298 | if tag is not None: |
---|
| 299 | handle, suffix = tag |
---|
| 300 | if handle is not None: |
---|
| 301 | if handle not in self.tag_handles: |
---|
| 302 | raise ParserError("while parsing a node", start_mark, |
---|
| 303 | "found undefined tag handle %r" % handle.encode('utf-8'), |
---|
| 304 | tag_mark) |
---|
| 305 | tag = self.tag_handles[handle]+suffix |
---|
| 306 | else: |
---|
| 307 | tag = suffix |
---|
| 308 | #if tag == u'!': |
---|
| 309 | # raise ParserError("while parsing a node", start_mark, |
---|
| 310 | # "found non-specific tag '!'", tag_mark, |
---|
| 311 | # "Please check 'http://pyyaml.org/wiki/YAMLNonSpecificTag' and share your opinion.") |
---|
| 312 | if start_mark is None: |
---|
| 313 | start_mark = end_mark = self.peek_token().start_mark |
---|
| 314 | event = None |
---|
| 315 | implicit = (tag is None or tag == u'!') |
---|
| 316 | if indentless_sequence and self.check_token(BlockEntryToken): |
---|
| 317 | end_mark = self.peek_token().end_mark |
---|
| 318 | event = SequenceStartEvent(anchor, tag, implicit, |
---|
| 319 | start_mark, end_mark) |
---|
| 320 | self.state = self.parse_indentless_sequence_entry |
---|
| 321 | else: |
---|
| 322 | if self.check_token(ScalarToken): |
---|
| 323 | token = self.get_token() |
---|
| 324 | end_mark = token.end_mark |
---|
| 325 | if (token.plain and tag is None) or tag == u'!': |
---|
| 326 | implicit = (True, False) |
---|
| 327 | elif tag is None: |
---|
| 328 | implicit = (False, True) |
---|
| 329 | else: |
---|
| 330 | implicit = (False, False) |
---|
| 331 | event = ScalarEvent(anchor, tag, implicit, token.value, |
---|
| 332 | start_mark, end_mark, style=token.style) |
---|
| 333 | self.state = self.states.pop() |
---|
| 334 | elif self.check_token(FlowSequenceStartToken): |
---|
| 335 | end_mark = self.peek_token().end_mark |
---|
| 336 | event = SequenceStartEvent(anchor, tag, implicit, |
---|
| 337 | start_mark, end_mark, flow_style=True) |
---|
| 338 | self.state = self.parse_flow_sequence_first_entry |
---|
| 339 | elif self.check_token(FlowMappingStartToken): |
---|
| 340 | end_mark = self.peek_token().end_mark |
---|
| 341 | event = MappingStartEvent(anchor, tag, implicit, |
---|
| 342 | start_mark, end_mark, flow_style=True) |
---|
| 343 | self.state = self.parse_flow_mapping_first_key |
---|
| 344 | elif block and self.check_token(BlockSequenceStartToken): |
---|
| 345 | end_mark = self.peek_token().start_mark |
---|
| 346 | event = SequenceStartEvent(anchor, tag, implicit, |
---|
| 347 | start_mark, end_mark, flow_style=False) |
---|
| 348 | self.state = self.parse_block_sequence_first_entry |
---|
| 349 | elif block and self.check_token(BlockMappingStartToken): |
---|
| 350 | end_mark = self.peek_token().start_mark |
---|
| 351 | event = MappingStartEvent(anchor, tag, implicit, |
---|
| 352 | start_mark, end_mark, flow_style=False) |
---|
| 353 | self.state = self.parse_block_mapping_first_key |
---|
| 354 | elif anchor is not None or tag is not None: |
---|
| 355 | # Empty scalars are allowed even if a tag or an anchor is |
---|
| 356 | # specified. |
---|
| 357 | event = ScalarEvent(anchor, tag, (implicit, False), u'', |
---|
| 358 | start_mark, end_mark) |
---|
| 359 | self.state = self.states.pop() |
---|
| 360 | else: |
---|
| 361 | if block: |
---|
| 362 | node = 'block' |
---|
| 363 | else: |
---|
| 364 | node = 'flow' |
---|
| 365 | token = self.peek_token() |
---|
| 366 | raise ParserError("while parsing a %s node" % node, start_mark, |
---|
| 367 | "expected the node content, but found %r" % token.id, |
---|
| 368 | token.start_mark) |
---|
| 369 | return event |
---|
| 370 | |
---|
| 371 | # block_sequence ::= BLOCK-SEQUENCE-START (BLOCK-ENTRY block_node?)* BLOCK-END |
---|
| 372 | |
---|
| 373 | def parse_block_sequence_first_entry(self): |
---|
| 374 | token = self.get_token() |
---|
| 375 | self.marks.append(token.start_mark) |
---|
| 376 | return self.parse_block_sequence_entry() |
---|
| 377 | |
---|
| 378 | def parse_block_sequence_entry(self): |
---|
| 379 | if self.check_token(BlockEntryToken): |
---|
| 380 | token = self.get_token() |
---|
| 381 | if not self.check_token(BlockEntryToken, BlockEndToken): |
---|
| 382 | self.states.append(self.parse_block_sequence_entry) |
---|
| 383 | return self.parse_block_node() |
---|
| 384 | else: |
---|
| 385 | self.state = self.parse_block_sequence_entry |
---|
| 386 | return self.process_empty_scalar(token.end_mark) |
---|
| 387 | if not self.check_token(BlockEndToken): |
---|
| 388 | token = self.peek_token() |
---|
| 389 | raise ParserError("while parsing a block collection", self.marks[-1], |
---|
| 390 | "expected <block end>, but found %r" % token.id, token.start_mark) |
---|
| 391 | token = self.get_token() |
---|
| 392 | event = SequenceEndEvent(token.start_mark, token.end_mark) |
---|
| 393 | self.state = self.states.pop() |
---|
| 394 | self.marks.pop() |
---|
| 395 | return event |
---|
| 396 | |
---|
| 397 | # indentless_sequence ::= (BLOCK-ENTRY block_node?)+ |
---|
| 398 | |
---|
| 399 | def parse_indentless_sequence_entry(self): |
---|
| 400 | if self.check_token(BlockEntryToken): |
---|
| 401 | token = self.get_token() |
---|
| 402 | if not self.check_token(BlockEntryToken, |
---|
| 403 | KeyToken, ValueToken, BlockEndToken): |
---|
| 404 | self.states.append(self.parse_indentless_sequence_entry) |
---|
| 405 | return self.parse_block_node() |
---|
| 406 | else: |
---|
| 407 | self.state = self.parse_indentless_sequence_entry |
---|
| 408 | return self.process_empty_scalar(token.end_mark) |
---|
| 409 | token = self.peek_token() |
---|
| 410 | event = SequenceEndEvent(token.start_mark, token.start_mark) |
---|
| 411 | self.state = self.states.pop() |
---|
| 412 | return event |
---|
| 413 | |
---|
| 414 | # block_mapping ::= BLOCK-MAPPING_START |
---|
| 415 | # ((KEY block_node_or_indentless_sequence?)? |
---|
| 416 | # (VALUE block_node_or_indentless_sequence?)?)* |
---|
| 417 | # BLOCK-END |
---|
| 418 | |
---|
| 419 | def parse_block_mapping_first_key(self): |
---|
| 420 | token = self.get_token() |
---|
| 421 | self.marks.append(token.start_mark) |
---|
| 422 | return self.parse_block_mapping_key() |
---|
| 423 | |
---|
| 424 | def parse_block_mapping_key(self): |
---|
| 425 | if self.check_token(KeyToken): |
---|
| 426 | token = self.get_token() |
---|
| 427 | if not self.check_token(KeyToken, ValueToken, BlockEndToken): |
---|
| 428 | self.states.append(self.parse_block_mapping_value) |
---|
| 429 | return self.parse_block_node_or_indentless_sequence() |
---|
| 430 | else: |
---|
| 431 | self.state = self.parse_block_mapping_value |
---|
| 432 | return self.process_empty_scalar(token.end_mark) |
---|
| 433 | if not self.check_token(BlockEndToken): |
---|
| 434 | token = self.peek_token() |
---|
| 435 | raise ParserError("while parsing a block mapping", self.marks[-1], |
---|
| 436 | "expected <block end>, but found %r" % token.id, token.start_mark) |
---|
| 437 | token = self.get_token() |
---|
| 438 | event = MappingEndEvent(token.start_mark, token.end_mark) |
---|
| 439 | self.state = self.states.pop() |
---|
| 440 | self.marks.pop() |
---|
| 441 | return event |
---|
| 442 | |
---|
| 443 | def parse_block_mapping_value(self): |
---|
| 444 | if self.check_token(ValueToken): |
---|
| 445 | token = self.get_token() |
---|
| 446 | if not self.check_token(KeyToken, ValueToken, BlockEndToken): |
---|
| 447 | self.states.append(self.parse_block_mapping_key) |
---|
| 448 | return self.parse_block_node_or_indentless_sequence() |
---|
| 449 | else: |
---|
| 450 | self.state = self.parse_block_mapping_key |
---|
| 451 | return self.process_empty_scalar(token.end_mark) |
---|
| 452 | else: |
---|
| 453 | self.state = self.parse_block_mapping_key |
---|
| 454 | token = self.peek_token() |
---|
| 455 | return self.process_empty_scalar(token.start_mark) |
---|
| 456 | |
---|
| 457 | # flow_sequence ::= FLOW-SEQUENCE-START |
---|
| 458 | # (flow_sequence_entry FLOW-ENTRY)* |
---|
| 459 | # flow_sequence_entry? |
---|
| 460 | # FLOW-SEQUENCE-END |
---|
| 461 | # flow_sequence_entry ::= flow_node | KEY flow_node? (VALUE flow_node?)? |
---|
| 462 | # |
---|
| 463 | # Note that while production rules for both flow_sequence_entry and |
---|
| 464 | # flow_mapping_entry are equal, their interpretations are different. |
---|
| 465 | # For `flow_sequence_entry`, the part `KEY flow_node? (VALUE flow_node?)?` |
---|
| 466 | # generate an inline mapping (set syntax). |
---|
| 467 | |
---|
| 468 | def parse_flow_sequence_first_entry(self): |
---|
| 469 | token = self.get_token() |
---|
| 470 | self.marks.append(token.start_mark) |
---|
| 471 | return self.parse_flow_sequence_entry(first=True) |
---|
| 472 | |
---|
| 473 | def parse_flow_sequence_entry(self, first=False): |
---|
| 474 | if not self.check_token(FlowSequenceEndToken): |
---|
| 475 | if not first: |
---|
| 476 | if self.check_token(FlowEntryToken): |
---|
| 477 | self.get_token() |
---|
| 478 | else: |
---|
| 479 | token = self.peek_token() |
---|
| 480 | raise ParserError("while parsing a flow sequence", self.marks[-1], |
---|
| 481 | "expected ',' or ']', but got %r" % token.id, token.start_mark) |
---|
| 482 | |
---|
| 483 | if self.check_token(KeyToken): |
---|
| 484 | token = self.peek_token() |
---|
| 485 | event = MappingStartEvent(None, None, True, |
---|
| 486 | token.start_mark, token.end_mark, |
---|
| 487 | flow_style=True) |
---|
| 488 | self.state = self.parse_flow_sequence_entry_mapping_key |
---|
| 489 | return event |
---|
| 490 | elif not self.check_token(FlowSequenceEndToken): |
---|
| 491 | self.states.append(self.parse_flow_sequence_entry) |
---|
| 492 | return self.parse_flow_node() |
---|
| 493 | token = self.get_token() |
---|
| 494 | event = SequenceEndEvent(token.start_mark, token.end_mark) |
---|
| 495 | self.state = self.states.pop() |
---|
| 496 | self.marks.pop() |
---|
| 497 | return event |
---|
| 498 | |
---|
| 499 | def parse_flow_sequence_entry_mapping_key(self): |
---|
| 500 | token = self.get_token() |
---|
| 501 | if not self.check_token(ValueToken, |
---|
| 502 | FlowEntryToken, FlowSequenceEndToken): |
---|
| 503 | self.states.append(self.parse_flow_sequence_entry_mapping_value) |
---|
| 504 | return self.parse_flow_node() |
---|
| 505 | else: |
---|
| 506 | self.state = self.parse_flow_sequence_entry_mapping_value |
---|
| 507 | return self.process_empty_scalar(token.end_mark) |
---|
| 508 | |
---|
| 509 | def parse_flow_sequence_entry_mapping_value(self): |
---|
| 510 | if self.check_token(ValueToken): |
---|
| 511 | token = self.get_token() |
---|
| 512 | if not self.check_token(FlowEntryToken, FlowSequenceEndToken): |
---|
| 513 | self.states.append(self.parse_flow_sequence_entry_mapping_end) |
---|
| 514 | return self.parse_flow_node() |
---|
| 515 | else: |
---|
| 516 | self.state = self.parse_flow_sequence_entry_mapping_end |
---|
| 517 | return self.process_empty_scalar(token.end_mark) |
---|
| 518 | else: |
---|
| 519 | self.state = self.parse_flow_sequence_entry_mapping_end |
---|
| 520 | token = self.peek_token() |
---|
| 521 | return self.process_empty_scalar(token.start_mark) |
---|
| 522 | |
---|
| 523 | def parse_flow_sequence_entry_mapping_end(self): |
---|
| 524 | self.state = self.parse_flow_sequence_entry |
---|
| 525 | token = self.peek_token() |
---|
| 526 | return MappingEndEvent(token.start_mark, token.start_mark) |
---|
| 527 | |
---|
| 528 | # flow_mapping ::= FLOW-MAPPING-START |
---|
| 529 | # (flow_mapping_entry FLOW-ENTRY)* |
---|
| 530 | # flow_mapping_entry? |
---|
| 531 | # FLOW-MAPPING-END |
---|
| 532 | # flow_mapping_entry ::= flow_node | KEY flow_node? (VALUE flow_node?)? |
---|
| 533 | |
---|
| 534 | def parse_flow_mapping_first_key(self): |
---|
| 535 | token = self.get_token() |
---|
| 536 | self.marks.append(token.start_mark) |
---|
| 537 | return self.parse_flow_mapping_key(first=True) |
---|
| 538 | |
---|
| 539 | def parse_flow_mapping_key(self, first=False): |
---|
| 540 | if not self.check_token(FlowMappingEndToken): |
---|
| 541 | if not first: |
---|
| 542 | if self.check_token(FlowEntryToken): |
---|
| 543 | self.get_token() |
---|
| 544 | else: |
---|
| 545 | token = self.peek_token() |
---|
| 546 | raise ParserError("while parsing a flow mapping", self.marks[-1], |
---|
| 547 | "expected ',' or '}', but got %r" % token.id, token.start_mark) |
---|
| 548 | if self.check_token(KeyToken): |
---|
| 549 | token = self.get_token() |
---|
| 550 | if not self.check_token(ValueToken, |
---|
| 551 | FlowEntryToken, FlowMappingEndToken): |
---|
| 552 | self.states.append(self.parse_flow_mapping_value) |
---|
| 553 | return self.parse_flow_node() |
---|
| 554 | else: |
---|
| 555 | self.state = self.parse_flow_mapping_value |
---|
| 556 | return self.process_empty_scalar(token.end_mark) |
---|
| 557 | elif not self.check_token(FlowMappingEndToken): |
---|
| 558 | self.states.append(self.parse_flow_mapping_empty_value) |
---|
| 559 | return self.parse_flow_node() |
---|
| 560 | token = self.get_token() |
---|
| 561 | event = MappingEndEvent(token.start_mark, token.end_mark) |
---|
| 562 | self.state = self.states.pop() |
---|
| 563 | self.marks.pop() |
---|
| 564 | return event |
---|
| 565 | |
---|
| 566 | def parse_flow_mapping_value(self): |
---|
| 567 | if self.check_token(ValueToken): |
---|
| 568 | token = self.get_token() |
---|
| 569 | if not self.check_token(FlowEntryToken, FlowMappingEndToken): |
---|
| 570 | self.states.append(self.parse_flow_mapping_key) |
---|
| 571 | return self.parse_flow_node() |
---|
| 572 | else: |
---|
| 573 | self.state = self.parse_flow_mapping_key |
---|
| 574 | return self.process_empty_scalar(token.end_mark) |
---|
| 575 | else: |
---|
| 576 | self.state = self.parse_flow_mapping_key |
---|
| 577 | token = self.peek_token() |
---|
| 578 | return self.process_empty_scalar(token.start_mark) |
---|
| 579 | |
---|
| 580 | def parse_flow_mapping_empty_value(self): |
---|
| 581 | self.state = self.parse_flow_mapping_key |
---|
| 582 | return self.process_empty_scalar(self.peek_token().start_mark) |
---|
| 583 | |
---|
| 584 | def process_empty_scalar(self, mark): |
---|
| 585 | return ScalarEvent(None, None, (True, False), u'', mark, mark) |
---|
| 586 | |
---|