strip_context

scripts.fix_tokenization.strip_context(new_context, old_context, start_offset, end_offset)[source]
>>> strip_context(["I", "'ve", "done"], ["I", "'ve", "done"], 1, 1)
["'ve"]
>>> strip_context(["I", "'ve,", "done"], ["I", "'ve", ",", "done"], 1, 1)
["'ve", ","]
>>> strip_context(["'ve", "done"], ["'", "ve", "done"], 0, 1)
["'", "ve"]
>>> strip_context(["I", "'ve,"], ["I", "'ve", ","], 1, 0)
["'ve", ","]
>>> strip_context(["can", "'t", "see"], ["ca", "n't", "see"], 1, 1)
["'t"]
>>> strip_context(["I", "can", "'t"], ["I", "ca", "n't"], 1, 1)
["can"]
>>> strip_context(["because", "somebody", "'d"], ["because", "somebody'd"], 1, 1)
[]
>>> strip_context(["somebody", "'d", "always"], ["somebody'd", "always"], 1, 1)
[]