About 10,000 deep learning papers have been written about "hard-coding priors about a specific task into a NN architecture works better than a lack of prior" -- but they're typically being passed as "architecture XYZ offers superior performance for [overly generic task category]"
-
-
Another way to inject prior knowledge is by adding to your network differentiable operations that you have identified as building blocks of the solution (e.g. if you know the solution involves cosines, you can add cos ops to your architecture).
-
And you say there are 10,000 papers (approx.) doing this?
- Show replies
New conversation -
-
-
Are you referring to the original bAbI paper, or a later one?
-
You can rank every bAbI-solving paper by how closely the solution architecture maps to the data-generation template. It's basically the same as ranking by accuracy. This is true to most tasks outside of perception -- bAbI is just an extreme example.
End of conversation
New conversation -
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.