{ "id":"eeeb2d01-8315-454e-b33f-3d6caa25db45", "title":"Lexical analysis [Wiki]", "authors":["Moeid Heidari","Denis Gorbunov"], "category": "Computer Science", "topic":"Compiler Engineering", "subTopic":"Lexical Analysis", "summary":"Lorem Ipsum is simply dummy text of the printing and typesetting industry...", "tags":["tag1","tag2","tag3"], "content":"# Lexical Analysis\n In computer science, lexical analysis, lexing or tokenization is the process of converting a sequence of characters (such as in a computer program or web page) into a sequence of lexical tokens (strings with an assigned and thus identified meaning). A program that performs lexical analysis may be termed a lexer, tokenizer,[1] or scanner, although scanner is also a term for the first stage of a lexer. A lexer is generally combined with a parser, which together analyze the syntax of programming languages, web pages, and so forth." }