Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Some plan #83

Open
3 tasks done
chengchingwen opened this issue Feb 10, 2022 · 5 comments
Open
3 tasks done

Some plan #83

chengchingwen opened this issue Feb 10, 2022 · 5 comments
Assignees

Comments

@chengchingwen
Copy link
Owner

chengchingwen commented Feb 10, 2022

Here are some stuff I'm going to rewrite for the new release:

feel free to add comments.

@chengchingwen chengchingwen self-assigned this Feb 10, 2022
@chengchingwen
Copy link
Owner Author

chengchingwen commented May 24, 2022

The new tokenizer api (using TextEncodeBase) is basically finished and included in the 0.1.16 release, though the gpt part is ignored for now. For the next step, I will be fixing the huggingface download issue with HuggingFaceApi.jl. Rewriting the attention layer might be breaking, so that would probably be the last one to do.

Some other issue that might also need to be tracked:

  • update gpt with the new api
  • Most documents are outdated
  • "time-to-first-plot" issue
  • Redesign (or get rid of) the Datasets api
  • Remove internal use of Stacks

@MNLubov
Copy link

MNLubov commented Jul 4, 2022

@chengchingwen
Peter, what is the approximate timeframe for implementing the model transfer from Huggingface?

@chengchingwen
Copy link
Owner Author

@MNLubov Are you looking for a specific model from HuggingFace? I'm trying to fix the huggingface module this month, so if everything goes well, it would be workable again before August.

Just to clarify, even if that huggingface module is fixed, it's still possible that we don't have the implementation for that model type (by model type, I mean something like bert, gpt2, t5 etc). So if you are looking for a model type that we don't have, please open another issue (and the timeline would be unknown for now

@MNLubov
Copy link

MNLubov commented Jul 4, 2022

@chengchingwen Thanks for the clarification. Currently I am testing different sentence-transformers from Huggingface to find the most suitable for my purposes. As a temporary solution, I use PyCall to find the most suitable one.
As far as I understand you have now bert, gpt and roberta implementations.

@chengchingwen
Copy link
Owner Author

@MNLubov Yes. I haven't investigate the sentence-transformers implementation, but it seem that it can also be done with normal huggingface interface. Like this one https://huggingface.co/sentence-transformers/all-MiniLM-L12-v2, it's a bert model, so it should be workable following the huggingface transformer usage in the readme after we fix the module.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants