Showing 204 of 204 total issues
Avoid deeply nested control flow statements. Open
Open
for sentence in wl_sentence_tokenization.wl_sentence_split(self.main, text_no_tags):
self.tokens_multilevel[-1].append([])
for sentence_seg in wl_sentence_tokenization.wl_sentence_seg_tokenize_tokens(
self.main,
Avoid deeply nested control flow statements. Open
Open
if re_match(search_term, dependency_relation.display_text(), flags = re_flags):
search_results.add(dependency_relation)
else:
Avoid deeply nested control flow statements. Open
Open
for sentence_seg in sentence:
for i, token in enumerate(sentence_seg):
if token in stop_words:
sentence_seg[i] = ''
else:
Avoid deeply nested control flow statements. Open
Open
for sentence in doc.sents:
htmls.append(spacy.displacy.render(
sentence,
style = 'dep',
minify = True,
Avoid deeply nested control flow statements. Open
Open
if not self.isRowHidden(row):
item = self.model().item(row, col)
val_cum += item.val
item.setText(f'{val_cum:.{precision_pcts}%}')
Avoid deeply nested control flow statements. Open
Open
if not self.isRowHidden(row):
item = self.model().item(row, col)
item.setText(f'{item.val:.{precision_decimals}}')
# Percentages
Avoid deeply nested control flow statements. Open
Open
for word in wl_word_tokenization.wl_word_tokenize_flat(main, tr, lang):
add_val_to_trs(trs_lexicon, word, vals)
else:
Avoid deeply nested control flow statements. Open
Open
if self.settings_tags == 'body_tag_settings' and tag_name == '*':
opening_tag_text = opening_tag_text.replace('*', _tr('wl_settings_files', 'TAG'))
closing_tag_text = self.model().item(row, 3).text().replace('*', _tr('wl_settings_files', 'TAG'))
preview.setText(opening_tag_text + _tr('wl_settings_files', 'token') + closing_tag_text)
else:
Avoid deeply nested control flow statements. Open
Open
if (para := para[tag_last_end:]):
tags_tokens = self.add_tags_splitting(para, tags_tokens)
# Add empty tags for untagged files
if not self.tagged:
Avoid deeply nested control flow statements. Open
Open
for token, lemma_search in set(zip(tokens, lemmas_search)):
if re_match(lemma_matched, lemma_search, flags = re_flags):
search_results.add(token)
Avoid deeply nested control flow statements. Open
Open
for node in nodes:
len_node = len(node)
for j, ngram in enumerate(wl_nlp_utils.ngrams(parallel_unit, len_node)):
if ngram == tuple(node):
Avoid deeply nested control flow statements. Open
Open
for k, ngram in enumerate(wl_nlp_utils.ngrams(tokens, len_search_term)):
if ngram == search_term:
points.append([x_start + k / text.num_tokens * len_tokens_total, y_start - j])
# Total
points.append([x_start_total + k, 0])
Avoid deeply nested control flow statements. Open
Open
if (
(
(
not settings['search_settings']['match_dependency_relations']
and (token in search_terms or head in search_terms)
Avoid deeply nested control flow statements. Open
Open
for j, collocate in enumerate(reversed(tokens_left)):
if wl_matching.check_context(
i, tokens,
context_settings = settings['search_settings']['context_settings'],
search_terms_incl = search_terms_incl,
Avoid deeply nested control flow statements. Open
Open
if ngram == tuple(search_term):
self.dialog.items_found.append([table, row, col])
Avoid deeply nested control flow statements. Open
Open
for j, collocate in enumerate(tokens_right):
if wl_matching.check_context(
i, tokens,
context_settings = settings['search_settings']['context_settings'],
search_terms_incl = search_terms_incl,
Avoid deeply nested control flow statements. Open
Open
for file in glob.glob(os.path.join(
self.settings_custom['general']['imp']['temp_files']['default_path'], '*.*'
)):
os.remove(file)
Avoid deeply nested control flow statements. Open
Open
if settings_limit_searching == _tr('wl_collocation_extractor', 'None'):
tokens_left = tokens[max(0, i + window_left) : max(0, i + window_right + 1)]
else:
# Span positions (Left)
for position in range(max(0, i + window_left), max(0, i + window_right + 1)):
Avoid deeply nested control flow statements. Open
Open
if ngram == search_term:
collocations_freqs_file_filtered[(node, collocate)] = freqs
Avoid deeply nested control flow statements. Open
Open
if _tr('wl_wordlist_generator', 'No language support') in tokens_syllabified:
self.set_item_err(i, 2, tokens_syllabified, alignment_hor = 'left')
else:
self.model().setItem(i, 2, wl_tables.Wl_Table_Item(tokens_syllabified))