This repository has been archived by the owner on Oct 19, 2024. It is now read-only.
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Moving all data generation projects to a single
datagen
project which carries all conda dependencies. Everything else can then be purely managed by Poetry and have cleaner dependency resolution. Gets rid of top levelenvironment.yaml
and poetry info to avoid confusion about how these are expected to be used. Closes #191 , closes #203 . Other salient changesinjection
library moved into newdatagen
trainer
library carries Torch GPU dependenciesOther changes to make in future PRs for simplification
WhiteningTransform
andbbhnet.data.distributions
submodules make it intoml4gw
, the remaining dataloader and injector subclasses as well as the glitch sampler should be moved intoprojects/sandbox/train
and thebbhnet.data
project should be removed altogether.bbhnet.architectures
should import theWhiteningTransform
fromml4gw
bbhnet.parallelize
altogether. We're much more conservative about how we submit futures now, so we don't need the defaultcancel_futures
functionality that theAsyncExecutor
class primarily serves for. For thedict
-basedas_completed
function, we can drop this in favor of using theFuture
as the key of the dictionary which will map to the corresponding metadata. We can then call a regularwait
onfutures.keys()
, thenpop
each key that gets returned asdone
.datagen.utils.injection.injection_waveforms
, we might consider dropping the tuplebackground
argument and the time array altogether in favor of accepting asample_rate
argument under the assumption that all of thesignal_times
have been normalized such thatt=0
represents the start ofbackground
. Alternatively, we can even dropsample_rate
and accept like asignal_offset
argument that has already done the conversion to number of samples.