Lines Matching full:tokenizer

43 std::string Parser::make_error(const Tokenizer& tokenizer,  in make_error()  argument
45 return std::to_string(tokenizer.GetCurrentLine()) + ": " + err; in make_error()
152 Tokenizer tokenizer(section.contents); in ProcessRequireBlock() local
153 tokenizer.SetCurrentLine(section.starting_line_number + 1); in ProcessRequireBlock()
155 for (auto token = tokenizer.NextToken(); !token->IsEOS(); in ProcessRequireBlock()
156 token = tokenizer.NextToken()) { in ProcessRequireBlock()
161 tokenizer, in ProcessRequireBlock()
169 token = tokenizer.NextToken(); in ProcessRequireBlock()
171 return Result(make_error(tokenizer, "Missing framebuffer format")); in ProcessRequireBlock()
177 make_error(tokenizer, "Failed to parse framebuffer format: " + in ProcessRequireBlock()
189 token = tokenizer.NextToken(); in ProcessRequireBlock()
191 return Result(make_error(tokenizer, "Missing depthStencil format")); in ProcessRequireBlock()
197 make_error(tokenizer, "Failed to parse depthstencil format: " + in ProcessRequireBlock()
221 token = tokenizer.NextToken(); in ProcessRequireBlock()
223 return Result(make_error(tokenizer, "Missing fence_timeout value")); in ProcessRequireBlock()
230 token = tokenizer.NextToken(); in ProcessRequireBlock()
233 tokenizer, "Missing width and height for fbsize command")); in ProcessRequireBlock()
237 make_error(tokenizer, "Invalid width for fbsize command")); in ProcessRequireBlock()
242 token = tokenizer.NextToken(); in ProcessRequireBlock()
245 make_error(tokenizer, "Missing height for fbsize command")); in ProcessRequireBlock()
249 make_error(tokenizer, "Invalid height for fbsize command")); in ProcessRequireBlock()
259 make_error(tokenizer, "Unknown feature or extension: " + str)); in ProcessRequireBlock()
265 token = tokenizer.NextToken(); in ProcessRequireBlock()
268 tokenizer, "Failed to parse requirements block: invalid token: " + in ProcessRequireBlock()
278 Tokenizer tokenizer(section.contents); in ProcessIndicesBlock() local
279 tokenizer.SetCurrentLine(section.starting_line_number); in ProcessIndicesBlock()
280 for (auto token = tokenizer.NextToken(); !token->IsEOS(); in ProcessIndicesBlock()
281 token = tokenizer.NextToken()) { in ProcessIndicesBlock()
286 return Result(make_error(tokenizer, "Invalid value in indices block: " + in ProcessIndicesBlock()
290 return Result(make_error(tokenizer, "Value too large in indices block: " + in ProcessIndicesBlock()
321 Tokenizer tokenizer(section.contents); in ProcessVertexDataBlock() local
322 tokenizer.SetCurrentLine(section.starting_line_number); in ProcessVertexDataBlock()
325 auto token = tokenizer.NextToken(); in ProcessVertexDataBlock()
327 token = tokenizer.NextToken(); in ProcessVertexDataBlock()
340 // Because of the way the tokenizer works we'll see a number then a string in ProcessVertexDataBlock()
344 make_error(tokenizer, "Unable to process vertex data header: " + in ProcessVertexDataBlock()
350 token = tokenizer.NextToken(); in ProcessVertexDataBlock()
353 make_error(tokenizer, "Unable to process vertex data header: " + in ProcessVertexDataBlock()
359 return Result(make_error(tokenizer, "Vertex data format too short: " + in ProcessVertexDataBlock()
366 make_error(tokenizer, "Invalid format in vertex data header: " + in ProcessVertexDataBlock()
375 token = tokenizer.NextToken(); in ProcessVertexDataBlock()
383 for (; !token->IsEOS(); token = tokenizer.NextToken()) { in ProcessVertexDataBlock()
395 make_error(tokenizer, "Invalid packed value in Vertex Data: " + in ProcessVertexDataBlock()
409 return Result(make_error(tokenizer, in ProcessVertexDataBlock()
424 return Result(make_error(tokenizer, "Invalid vertex data value: " + in ProcessVertexDataBlock()
429 token = tokenizer.NextToken(); in ProcessVertexDataBlock()