Glossary

Tokenization

The process of splitting text into individual units (tokens), typically words or sub-word units, as a prerequisite for most computational text analysis.

Learn More

Related Terms