You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: official/nlp/bert/README.md
+17Lines changed: 17 additions & 0 deletions
Original file line number
Diff line number
Diff line change
@@ -129,6 +129,23 @@ which is essentially branched from [BERT research repo](https://github.com/googl
129
129
to get processed pre-training data and it adapts to TF2 symbols and python3
130
130
compatibility.
131
131
132
+
Running the pre-training script requires an input and output directory, as well as a vocab file. Note that max_seq_length will need to match the sequence length parameter you specify when you run pre-training.
133
+
134
+
Example shell script to call create_pretraining_data.py
0 commit comments