Skip to content
This repository was archived by the owner on Nov 22, 2022. It is now read-only.

Commit 11f9088

Browse files
rutyrinottfacebook-github-bot
authored andcommitted
combine training data from multiple sources (#953)
Summary: Pull Request resolved: #953 Add data object that allows combining multiple data types during training and evaluation. Useful when training data is split across different files/partitions, and user wants to explore different combinations Reviewed By: borguz Differential Revision: D17143117 fbshipit-source-id: 881e8288359658826600408b72cf3e6887089630
1 parent 1a66bcd commit 11f9088

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

pytext/data/batch_sampler.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -154,14 +154,14 @@ class RandomizedBatchSampler(BaseBatchSampler):
154154
"""
155155
This sampler takes in a dictionary of iterators and returns batches according
156156
to the specified probabilities by `unnormalized_iterator_probs`. We cycle through
157-
the iterators (restarting any that "run out") indefinitely. Set epoch_size in
158-
Data.Config.
157+
the iterators (restarting any that "run out") indefinitely. Set batches_per_epoch
158+
in Trainer.Config.
159159
160160
Example:
161161
162162
Iterator A: [A, B, C, D], Iterator B: [a, b]
163163
164-
epoch_size = 3, unnormalized_iterator_probs = {"A": 0, "B": 1}
164+
batches_per_epoch = 3, unnormalized_iterator_probs = {"A": 0, "B": 1}
165165
Epoch 1 = [a, b, a]
166166
Epoch 2 = [b, a, b]
167167

0 commit comments

Comments
 (0)