You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This basically seems to work, although there are a few issues. One is that it's hard to interpret the merge list in the json file when I save it - which I suspect it related to the main issue, which is that I can't load the tokenizer after training it.
I get an error: Exception: data did not match any variant of untagged enum ModelWrapper at line 11182 column 3
It seems like it is related to: #566
I followed to: #909
which hasn't been merged, and suggests that merging with spaces probably isn't supported yet (without this PR). But things largely work up until the point where I try to load the tokenizer, which is strange.
Does anyone have any suggestions for getting around this, or an alternative approach to do the same thing?
The text was updated successfully, but these errors were encountered:
My use case is that I have a corpus of instructions, e.g. taking a small snapshot:
Each line is a separate "instruction", and I want to allow merges across the whole instruction (including whitespaces). My first attempt at this is:
This basically seems to work, although there are a few issues. One is that it's hard to interpret the merge list in the json file when I save it - which I suspect it related to the main issue, which is that I can't load the tokenizer after training it.
When I run:
I get an error:
Exception: data did not match any variant of untagged enum ModelWrapper at line 11182 column 3
It seems like it is related to: #566
I followed to: #909
which hasn't been merged, and suggests that merging with spaces probably isn't supported yet (without this PR). But things largely work up until the point where I try to load the tokenizer, which is strange.
Does anyone have any suggestions for getting around this, or an alternative approach to do the same thing?
The text was updated successfully, but these errors were encountered: