-
Notifications
You must be signed in to change notification settings - Fork 681
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
add transformer support?(matmul, layernorm) #216
Comments
Hello, this is very convenient repo to convert pytorch model to tensorrt, but it likely support cv well, support nlp poolly,I try to convert gpt2 model to tensorRT, but there are two operations that transformers(nlp very popular model, it is the base block to build gpt2, bert) needs don't support, include matmul, layernorm, c++ api support it, consider support it in python api? I am very expected to deploy gpt2 in tensorRT. |
I convert gpt2 model to tensorRT successfully use this repo and some operation I write by myself, and the speed has been increased three times, thanks this good repo. |
Can you share you impl of matmul, plz? |
Just use addMatrixMultiply() which is defined here |
can you share your ways |
@bigprince97 can you shart your code? thanks! |
Sorry to see the reply now, I write code last year, it can't run because of the update of the repo. fastertransformer supports gpt2 acceleration,It has pytorch interface, you can try. |
for anyone looking for the
|
No description provided.
The text was updated successfully, but these errors were encountered: