This looks like it has very good potential, but I can't seem to use any other activation function other than sigmoid, which my reinforcement learning problem requires more robust activation functions like ReLU/Leaky ReLU. Is there a way to change the derivatives in the code to accommodate different activation functions? Also, will this library gain a built-in way to check loss as a convenience for programmers using this?