Commit graph

1 commit

Author SHA1 Message Date
Pedro Rodriguez 374409fa3b [WIP] Changes for training entropy model and correcting attention in local models
Some checks are pending
Lint with Black / lint (push) Waiting to run
Lint with isort / lint (push) Waiting to run
Summary:

- Refactor local model configs to be separate and clearer
- Add attention arguments and correct which attention is used in local models
- Preparation for being able to have an entropy train script
- Fix failing unit tests

Test Plan:
2025-01-17 01:02:27 +00:00
Renamed from bytelatent/model/transformer.py (Browse further)