Thanks to visit codestin.com
Credit goes to github.com

Skip to content

Conversation

@jeffkinnison
Copy link
Contributor

LLMEncoder fine-tuning is currently blocked by two issues:

This PR adds fixes for both issues:

  • LLMEncoder output is now explicitly cast to torch.float32
  • The final layer of the base model is always frozen in LLMEncoder.__init__

@github-actions
Copy link

Unit Test Results

  6 files  ±0    6 suites  ±0   14m 27s ⏱️ +16s
12 tests ±0    9 ✔️ ±0    3 💤 ±0  0 ±0 
60 runs  ±0  42 ✔️ ±0  18 💤 ±0  0 ±0 

Results for commit 6721dfc. ± Comparison against base commit ea50811.

@jeffkinnison jeffkinnison merged commit 27c6079 into master Jan 19, 2024
@jeffkinnison jeffkinnison deleted the llm-encoder-dtype-mismatch branch January 19, 2024 18:36
vijayi1 pushed a commit to vijayi1/ludwig that referenced this pull request Jan 23, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants