Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About the print message #25

Open
lucasjinreal opened this issue Jul 21, 2023 · 2 comments
Open

About the print message #25

lucasjinreal opened this issue Jul 21, 2023 · 2 comments

Comments

@lucasjinreal
Copy link

Hi, the printed message I can not understand.

I set

def __init__(
        self, dim, ratio, max_position_embeddings=2048, base=10000, device=None
    ):

ratio=2, max_position_embeddings=1024

Since my GPU can not fit minimal 2048, so, I thought it was expanding the context from 1024 to 2048

But I got print like this:

Condensing Positional embeddings from 4096 to 2048

Which I don't understand?

@DachengLi1
Copy link
Owner

@lucasjinreal Thanks for the feedback! I think you are right, setting these should print from 2048 to 1024.

Can you print out the actual max_position_embeddings inside the function? I think huggingface may overwrite it during run-time (i.e. not using the 1024 you give)

@lucasjinreal
Copy link
Author

@DachengLi1 I just expose max_position_embeeding as a param in monkey patch, not sure what happened, but if so, if it was overwritten, then my actually training are 4096? (not 2048 as my expected).
But, my minial length set to 1024

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants