Source code for ASBC.queryParser

import re

[docs]def tokenize(string): """Parse query string for ngram into token objects Parameters ---------- string : str Query string with each token enclosed in a pair of square brackets. In each token, the tag ``word`` and ``pos`` could be given as ``[word="他們" pos="N.*"]``. To search with regex in ``word``, append ``.regex`` to ``word``: ``[word.regex="們$" pos="N.*"]``. ``pos`` by default uses regex search. Returns ------- list A list of token objects (dictionaries), with each dictionary representing the token in the query string (i.e. token enclosed in the brackets). Each token has three key-value pairs: - `tk`: ``str``. The pattern of the word to search for. - `tk.regex`: ``bool``. Whether to use regex search with word. - `pos`: ``str``. The pattern of the pos tag to search for. """ # Deal with single exact match of token if string.find("[") == -1: return [{ 'tk': string, 'pos': None, 'tk.regex': False, }] # Scan through the string to find matching brackets tokens = [] openPos =[] depth = 0 for i, char in enumerate(string): if char == '[': openPos.append(i) depth += 1 if char == ']': start = openPos.pop() depth -= 1 tokens.append({ 'start': start, 'end': i, 'inside': string[start+1:i], 'depth': depth }) # Get matching brackets at first depth level tk_pat = re.compile('''word=['"]([^'"]+)['"]''') pos_pat = re.compile('''pos=['"]([^'" ]+)['"]''') tkRegEx_pat = re.compile('''word.regex=['"]([^'"]+)['"]''') output = [] for tk in tokens: if tk['depth'] == 0: token = tk_pat.findall(tk['inside']) tkRegEx = tkRegEx_pat.findall(tk['inside']) token = tkRegEx if tkRegEx else token pos = pos_pat.findall(tk['inside']) output.append({ 'tk': token[0] if len(token) > 0 else None, 'pos': pos[0] if len(pos) > 0 else None, 'tk.regex': True if tkRegEx else False, }) return output
#%%
[docs]def querySpecificity(queryObj={'tk': '^我們$', 'pos': 'N%', 'tk.regex': True}): """Score a token object for specificity. Parameters ---------- queryObj : dict A token object in a list returned by :py:func:`.tokenize`. Returns ------- float A point indicating the specificity of the token. Higher score means the token is more specific and may result in fewer query results in the corpus. This point is used to determine the seed token of an ngram to search in the corpus (to boost performance). """ status = { 'token': { 'has_regEx': False, 'zh_len': 0 }, 'pos': { 'has_wildcard': False, 'tag_len': 0, } } #-------- Check token pattern --------# # List of regEx metacharacters indicating specific pattern regEx_meta = ['^', '$', '[', ']', '?' '{', '}', '(', ')', '|'] if queryObj['tk.regex'] and \ set(queryObj['tk']).intersection(regEx_meta): status['token']['has_regEx'] = True # Check chinese character if queryObj['tk'] is not None: for char in queryObj['tk']: if char > u'\u4e00' and char < u'\u9fff': status['token']['zh_len'] += 1 #------ Check pos tag pattern --------# if queryObj['pos'] is not None: if queryObj['pos'].find('%') != -1: status['pos']['has_wildcard'] = True for char in queryObj['pos']: if re.match('[A-Za-z]', char): status['pos']['tag_len'] += 1 return 1.2 * status['token']['zh_len'] + status['token']['has_regEx'] + \ 0.5 * status['pos']['tag_len'] - 0.2 * status['pos']['has_wildcard']