Concept

Unix Tools for Crude Tokenization and Normalization

Unix commands such as tr, sort, and uniq can be used for simple normalization, tokenization, and frequency computation.

0

1

Updated 2021-09-19

Tags

Data Science