Tokenizing a French text

Report a typo

To go further, manually open an IDE on your computer and copy the output to the website.

Let's work with French. To install French, use the following code:

import stanza
stanza.download('fr')
nlp = stanza.Pipeline(lang="fr", processors="tokenize, pos, lemma, depparse, ner")

You have a sentence in French. Print out all tokens of that sentence in one list.

Example: ['token', 'token', ',', 'token', 'token', '.']

Use Stanza 1.4.0. You can install it with this line:

!pip install stanza==1.4.0
Write code in your IDE to process the text file and display the results below
___

Create a free account to access the full topic