-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Incorrect check on concat original start index #2717
Merged
mravanelli
merged 1 commit into
speechbrain:develop
from
gfdb:concat-original-index-check-bug
Oct 11, 2024
Merged
Incorrect check on concat original start index #2717
mravanelli
merged 1 commit into
speechbrain:develop
from
gfdb:concat-original-index-check-bug
Oct 11, 2024
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Hi @gfdb, thank you for the fix. |
mravanelli
approved these changes
Oct 11, 2024
naspert
added a commit
to naspert/speechbrain
that referenced
this pull request
Oct 29, 2024
* data prep scripts update * iterate over utterances * without parallel map * parallel map -> so fast omfg * gigaspeech data prep done * speechcolab extra dep if one must download gigaspeech * create ASR CTC folder * base yaml + update data prep to better reflect potential different naming for csvs * update recipe * update recipe to be compliant with gigaspeech csv * add transformers dep * convert opus to wav * recipe --debug mode works. * typo GRABAGE_UTTERANCE_TAGS -> GARBAGE_UTTERANCE_TAGS * tmp DL file * update DL FILE * add DL file in ASR/CTC * update extra_requirements.txt * add support of savedir within Pretrained subclasses * add wbs requirements * webdataset * remove print * tmp files webdataset * verbosity + metada.json * letzo now label_encoder can actually train + the recipe seems to work. * remove wbs * DL info * HF DL support * remove webdataset as it sucks :p * name * ngram commands * whisper baseline * fix HF * pre-commit + sentencepiece char * remove csv * Add quirks.py, move global PyTorch config and GPU workarounds there * Add support for SB_DISABLE_QUIRKS environment variable * Fetch rework: make savedir optional * bunch of updates to make it run * no download script * fix precommit * fix precommit * readmes * readmes * readmes * readmes * doc update * CI god not happy, make CI god happy * why you here little encoder * adding a tranduscer streaming recipe, because why not * add test for transducer * works better when me not stupid * fix yaml * update req * add warning for cache dir * add warning for cache dir * enable multiprocessing * Minor cleanups to fetching * Change default behavior of inference to not create savedir if not specified * allow data prep without ddp * fix tests * smoll readme update * fix review comments * fixed concat_start_index check (speechbrain#2717) * Ensure adapted models save their parameters (speechbrain#2716) Co-authored-by: Parcollet Titouan <parcollet.titouan@gmail.com> * wtf * update doc * more documentation on storage * missing arg * a bit of logs * new schedulers * new schedulers * Fixes speechbrain#2656: Remove EOS from SoundChoice * fix my stupidity * Update non-HF code path for new preprocessing code in GigaSpeech * Fix CSV path for non-HF Gigaspeech * Fix formatting * Kmeans fix (speechbrain#2642) * fix kmeans bug * fix final batch * fix chuncksize * fix * fix * fix precommit * fix doxstrin inconsistency * fix precommit * fix doc string --------- Co-authored-by: Mirco Ravanelli <mirco.ravanelli@gmail.com> * add call on start of fit_batch fn * Update core.py Fix old commit * Update core.py * Fix preprocess_text example * Fix guess_source docstring with up-to-date info * Also remove default savedir from Pretrained * Fix function name for log_applied_quirks * wip audiomnist+gt * Revert "fix normalization for LFB" This reverts commit 3fd0330. * audiomnist classification setup * fix config * add missing file * update dataset load/training * remove unnecessary params * remove sort * remove unnecessary code * fix paths * fix loss computation * add missing flatten * print summary * Explain quirks in docs/experiment.md * ok stupid linter check that hates intentional leading spaces in markdown * add citing in README * add code to pad all wavs to the same length * fix pad call * fix error computation * fix error computation * Make `collect_in` optional for `Pretrainer`, disable it by default * Change more defaults to `savedir=None` and `fetch_strategy=SYMLINK` Since the SYMLINK strategy falls back to NO_LINK whenever `savedir is None`, it makes sense to switch more things to default to `savedir=None`. Should the `savedir` explicitly be set by the user, past behavior is preserved (defaulting to symlinks). * move flatten in audionet * Fix GS transducer test prediction decoding? * fix data prep logic and paths * Actually fix GS transducer test prediction decoding * Remove punctuation filtering that is handled elsewhere * HuggingFance * fix skip data prep logic * add original audionet feature extraction * fix pooling for audionet feature extraction * fix audionet shape + remove input norm * try data augmentation * add missing refs * - rework AudioNet to have optional pooling - use official AudioMNIST train/test/valid splits * fix typo in url * update audionet hparams * update audionet custom hparams * update audionet custom hparams * Updated warning for load_collected * Add results and notices for results for GigaSpeech transducer & wavlm * english hard * update audionet custom hparams * fix doc + pre-commit clean * fix code examples * fix consistency tests * fix pre commit * remove config * fix docstring for LFB * fix docstring for GammatoneConv1D --------- Co-authored-by: Adel Moumen <adelmoumen.pro@gmail.com> Co-authored-by: Adel Moumen <88119391+Adel-Moumen@users.noreply.github.com> Co-authored-by: asu <sdelang@sdelang.fr> Co-authored-by: TParcollet <parcollet.titouan@gmail.com> Co-authored-by: Peter Plantinga <plantinga.peter@proton.me> Co-authored-by: gianfranco <62777451+gfdb@users.noreply.github.com> Co-authored-by: Peter Plantinga <plantinga.peter@protonmail.com> Co-authored-by: Titouan Parcollet/Embedded AI /SRUK/Engineer/Samsung Electronics <t.parcollet@sruk-ccn4.eu.corp.samsungelectronics.net> Co-authored-by: flexthink <flexthink@users.noreply.github.com> Co-authored-by: Pooneh Mousavi <moosavi.pooneh@gmail.com> Co-authored-by: Mirco Ravanelli <mirco.ravanelli@gmail.com>
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
What does this PR do?
In the
Augmenter
class we allow the users to concatenate the original signals (the input batch) after augmentation. Furthermore, we allow them to select a subset of the original batch to concatenate viaconcat_start_index
andconcat_end_index
. This PR fixes a bug in the concatenation validation logic where we check to see if theconcat_start_index
exceeds the size of the input batch before performing the concatenation. Instead of checking the batch size of the original signalx_original.shape[0]
, we are currently checking the batch size of the subset of the input that the user has selected for augmentationx.shape[0]
which got changed when we didx = x[self.augment_start_index : self.augment_end_index_batch]
. This will skip the concatenation of the original signal when the batch dimension of the slice the user selects for augmentation is less than or equal toconcat_start_index
. These two have nothing to do with each other, this is a mistake.Before submitting
PR review
Reviewer checklist