-
Notifications
You must be signed in to change notification settings - Fork 34
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Use of causal models for generation #82
Comments
Hi @dipankarsrirag, thanks for your kind words. AnglE supports bi-directional LLMs. If you want to train AnglE embedding with bi-directional LLMs, you can refer to this documentation, in Examples/b.LLM-based If you just want to test the prompt with biLLM, you can directly use our BiLLM toolkit: https://github.com/WhereIsAI/BiLLM. It is compatible with huggingface transformers. |
Hi @SeanLee97, thanks for the quick reply. I have been working with AnglE for the past few hours now. Just need a clarification:
|
hi @dipankarsrirag, here are the answers to the questions:
|
This is an amazing work. I have been working on something that would require me to evaluate the generated outputs of models like Mistral, using a prompt like:
"Fill the [MASK] token in the sentence. Generate a single output."
Now earlier, I would simply instruction fine-tune a Mistral Model. But I would like to explore the possibility of using these models with a bi-directional attention.
I see that the library allows me to access the
backbone
model underneath. But it is not clear to me if this model has the bi-directional attention. Can you please clarify this? If it does, I could simply use thebackbone.generate()
function for my purpose.Thanks in advance!
The text was updated successfully, but these errors were encountered: