Incorporating neural networks in the material #4
Replies: 2 comments 1 reply
-
I think the general idea sounds laudable, but I see some challenges. In my view, a basic problem is that state-of-the-art does not necessarily mean pedagogical, and current or future state-of-the-art models might not work particularly well on manageable data sizes. Moreover, even a moderately ok performing neural network might not be very useful for understanding the phenomenon or the problem, but would still just be a bunch of layers and a loss function. I don't know which direction you want to take the book on the long run, but I would seriously consider
Also, I have no idea how to do this in a practical way, and I'm not familiar with the way how the current code execution of the book platform even works (i.e., is it locally when you read the book, locally when you compile the repository as a book for viewing, or on some server in either of the two cases?). |
Beta Was this translation helpful? Give feedback.
-
Here is a simple GRU-based speech enhancement module that could fit the format of NN modules that we want to have in terms of complexity and pedagogy. https://jmvalin.ca/demo/rnnoise/ |
Beta Was this translation helpful? Give feedback.
-
I'd like to start a major improvement task of importing baselines of neural networks in all applicable sections. Before starting, I'd like to have opinions on how to best include these models.
Some of my own preferences:
The main consequence of these preferences is that I think that each neural model should be in a separate repository. The idea is that the notebooks in the book would then load the external, pre-trained model for demonstrations. This leads to a follow-up question; should we have a single repo for all models in the book, or have a single repo for every model? I'm slightly in favour of the latter approach since this would make it easier to track library requirements and when the state-of-the-art changes, we need to keep track only of the current ones, not all of history.
Another question is whether we should primarily use our own implementations, or try to incorporate existing pre-trained models.
Especially hoping for comments from @orasanen
Beta Was this translation helpful? Give feedback.
All reactions