Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Feature/8mi trainer #834

Open
wants to merge 1 commit into
base: master
Choose a base branch
from
Open

Feature/8mi trainer #834

wants to merge 1 commit into from

Conversation

dpressel
Copy link
Collaborator

  • Adds an 8mi primitive called trainer to handle device specifics
  • Allow loaders to use "native" framework (ie tf's Dataset and pytorch's DataLoader) for batch delivery
  • Refactors TF eager mode to use only "native"
  • Creates custom sequence wrapper to allow pre-batching in TF, no need for custom data loader
  • Creates experimental addon called mead3_pytorch, supports future (distributed) training using 8mi via fit_func=mead3
  • Gets rid of dependency on six
  • Fixes PyTorch tagger handling of untrimmed data in neg_log_loss

@dpressel dpressel force-pushed the feature/8mi-trainer branch 2 times, most recently from c70b0b2 to a658ea8 Compare July 14, 2021 22:30
@dpressel dpressel force-pushed the feature/8mi-trainer branch from 3101a57 to b2b31f9 Compare August 15, 2021 14:50
WIP

WIP: remap tagger and classifier onto Trainer

oops fix issue with span F1 aggregation in Trainer

adds support for other loss functions like KLDiv

this is useful for cases like distillation where we
can have soft targets.

pass kwargs into target

use forward function

option whether to rm wrapper

support overriding the train target

This should fix multiworker mismatch on reload

feelgood types

fix first batch accum

allow no early stopping

global_step fix, clean examples, factor up

more cleanup

fix includes in addon

rm dist code outside 8mi trainer, WIP dataset

use native loaders via mead

pseudo fix for showing examples

fix default and backend arg in paired reader

bye six + tmp working non-native LM loader

add backend option

LM is TF native

fix test

remove and simplify tf trainers and fix trim issue

be a little tricky with TF native

we cant switch it on with TF 1.x

.

explore more modular refactoring

WIP

WIP: remap tagger and classifier onto Trainer

oops fix issue with span F1 aggregation in Trainer

adds support for other loss functions like KLDiv

this is useful for cases like distillation where we
can have soft targets.

pass kwargs into target

use forward function

option whether to rm wrapper

support overriding the train target

This should fix multiworker mismatch on reload

feelgood types

fix first batch accum

allow no early stopping

global_step fix, clean examples, factor up

more cleanup

fix includes in addon

rm dist code outside 8mi trainer, WIP dataset

use native loaders via mead

pseudo fix for showing examples

fix default and backend arg in paired reader

bye six + tmp working non-native LM loader

add backend option

LM is TF native

fix test

remove and simplify tf trainers and fix trim issue

.
@dpressel dpressel force-pushed the feature/8mi-trainer branch from b2b31f9 to 0432595 Compare August 15, 2021 14:58
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant