Skip to content

[chat] use official transformers and fix some issues#4117

Merged
ver217 merged 21 commits intohpcaitech:mainfrom
cwher:the-newer-the-better
Jul 4, 2023
Merged

[chat] use official transformers and fix some issues#4117
ver217 merged 21 commits intohpcaitech:mainfrom
cwher:the-newer-the-better

Conversation

@cwher
Copy link
Copy Markdown
Contributor

@cwher cwher commented Jun 29, 2023

📌 Checklist before creating the PR

  • I have created an issue for this PR for traceability
  • The title follows the standard format: [doc/gemini/tensor/...]: A concise description
  • I have added relevant tags if possible for us to better distinguish different PRs

🚨 Issue number

Link this PR to your issue with words like fixed to automatically close the linked issue upon merge

e.g. fixed #1234, closed #1234, resolved #1234

Close #4116 & #3639, #3389 & #3123, #3192.

📝 What does this PR do?

Summarize your work here.
if you have any plots/diagrams/screenshots/tables, please attach them here.

💥 Checklist before requesting a review

  • I have linked my PR to an issue (instruction)
  • My issue clearly describes the problem/feature/proposal, with diagrams/charts/table/code if possible
  • I have performed a self-review of my code
  • I have added thorough tests.
  • I have added docstrings for all the functions/methods I implemented

⭐️ Do you enjoy contributing to Colossal-AI?

  • 🌝 Yes, I do.
  • 🌚 No, I don't.

Tell us more if you don't enjoy contributing to Colossal-AI.

@cwher
Copy link
Copy Markdown
Contributor Author

cwher commented Jun 29, 2023

⚠️ This PR is not done yet.

The effect of removing prepare_llama_tokenizer_and_embedding hasn't been carefully examined.

@cwher
Copy link
Copy Markdown
Contributor Author

cwher commented Jun 29, 2023

Commit 7f7ecba solves #3123 and #3192.

@cwher cwher force-pushed the the-newer-the-better branch from 7f7ecba to 3ed6841 Compare June 29, 2023 10:23
@cwher cwher force-pushed the the-newer-the-better branch from 55d730b to baadbb1 Compare July 3, 2023 06:42
@cwher cwher changed the title [chat] use official transformers and fix some errors [chat] use official transformers and fix some issues Jul 3, 2023
@cwher cwher requested a review from ver217 July 3, 2023 10:19
Comment thread .github/workflows/run_chatgpt_examples.yml Outdated
Comment thread applications/Chat/examples/community/peft/train_peft_sft.py
Comment thread applications/Chat/examples/train_prompts.py
Comment thread applications/Chat/examples/train_sft.py
@ver217 ver217 merged commit 3d8d5d0 into hpcaitech:main Jul 4, 2023
ver217 pushed a commit to ver217/ColossalAI that referenced this pull request Jul 13, 2023
* feat: remove on_learn_epoch fn as not used

* revert: add _on_learn_epoch fn

* feat: remove NaiveStrategy

* test: update train_prompts tests

* fix: remove prepare_llama_tokenizer_and_embedding

* test: add lora arg

* feat: remove roberta support in train_prompts due to runtime errs

* feat: remove deberta & roberta in rm as not used

* test: remove deberta and roberta tests

* feat: remove deberta and roberta models as not used

* fix: remove calls to roberta

* fix: remove prepare_llama_tokenizer_and_embedding

* chore: update transformers version

* docs: update transformers version

* fix: fix actor inference

* fix: fix ci

* feat: change llama pad token to unk

* revert: revert ddp setup_distributed

* fix: change llama pad token to unk

* revert: undo unnecessary changes

* fix: use pip to install transformers
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

[chat] use official transformers and fix some errors

2 participants