non_flash branch #23
Closed
nassim12sol
started this conversation in
General
Replies: 1 comment
-
Hi, you can use it! We haven't thoroughly tested it, but representation embeddings are roughly the same as with the Flash-Attention-version. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi I have a question ,
Can i use the non_flash branch with the pretrained model , and would it provide the same results ?
Beta Was this translation helpful? Give feedback.
All reactions