verne.freenode.net changed the topic of #mlpack to: http://www.mlpack.org/ -- We don't respond instantly... but we will respond. Give it a few minutes. Or hours. -- Channel logs: http://www.mlpack.org/irc/
vivekp has quit [Ping timeout: 276 seconds]
vivekp has joined #mlpack
sumedhghaisas has quit [Read error: Connection reset by peer]
sumedhghaisas has joined #mlpack
govg has quit [Ping timeout: 240 seconds]
manthan has joined #mlpack
govg has joined #mlpack
witness has joined #mlpack
sulan_ has joined #mlpack
wenhao has joined #mlpack
ImQ009 has joined #mlpack
witness has quit [Quit: Connection closed for inactivity]
sulan_ has quit [Quit: Leaving]
rajeshdm9 has joined #mlpack
rajeshdm9 has quit [Client Quit]
sumedhghaisas has quit [Read error: Connection reset by peer]
sumedhghaisas has joined #mlpack
sumedhghaisas has quit [Read error: Connection reset by peer]
sumedhghaisas has joined #mlpack
sumedhghaisas2 has joined #mlpack
sumedhghaisas has quit [Ping timeout: 276 seconds]
Atharva has joined #mlpack
csoni has joined #mlpack
zoq_ has joined #mlpack
vpal has joined #mlpack
vpal is now known as vivekp
csoni has quit [Read error: Connection reset by peer]
zoq_ is now known as zoq
dmatt has joined #mlpack
Atharva has quit [Quit: Connection closed for inactivity]
dmatt has quit [Remote host closed the connection]
sumedhghaisas2 has quit [Read error: Connection reset by peer]
sumedhghaisas has joined #mlpack
govg has quit [Quit: leaving]
Atharva has joined #mlpack
< Atharva> rcurtin: zoq: what irc client do you use to stay connected forever? Do you use it on phone or pc?
< rcurtin> Atharva: I use irssi in a GNU screen session on a server that I host (it's the same server that hosts mlpack.org)
< rcurtin> when I read messages, I simply connect from whatever computer I am using with ssh and resume the screen session
< rcurtin> I think it is an unusual setup but it works for me :)
< Atharva> Oh, okay, I don’t have a server for that kind of a setup. I think I would have to just search the net for a good client.
< rcurtin> I know there are IRC bouncers out there... I want to say one of these is called 'matrix'? but I am not certain
< Atharva> Yeah, I have heard about the bouncers, not quite sure what they do. I will check them out. There are cloud based clients which keep you online all the time but they charge on a per month basis and are quite expensive.
< rcurtin> yeah, I thought there was at least one that was free
< rcurtin> but even if you don't have a way to have a client always in the room, it is logged to http://www.mlpack.org/irc/, so you can always keep an eye on that :)
< Atharva> That is extremely useful, I always use the logs.
< Atharva> I had this another doubt, how do I build everything else except the test framework. I am trying some changes in the mlpack ann codebase and some tests fail to compile because of that.
< rcurtin> you can configure cmake with -DBUILD_TESTS=OFF
< rcurtin> and then when you type 'make', by default, the tests will not be built (you should still be able to type 'make mlpack_test' if you want the tests)
< Atharva> It’s still failing, I think I will have to check the changes I made. How do I compile just the ANN module?
< rcurtin> well, so this one is a little bit tricky
< rcurtin> the ANN code doesn't actually compile into anything because it is header-only
< rcurtin> this is the case with a lot of code in mlpack (but not all of it---any .cpp files in src/mlpack/core and src/mlpack/methods get compiled into libmlpack.so)
< rcurtin> so the only way it gets compiled into something is either in the tests in src/mlpack/tests/ or in the bindings found in src/mlpack/methods/*/*_main.cpp
sumedhghaisas has quit [Ping timeout: 276 seconds]
< rcurtin> ok... I think that I have mlpack pypi packages compiling successfully. once I verify they are working right I'll upload the scripts into the jenkins-conf repository
wenhao has quit [Ping timeout: 260 seconds]
vivekp has quit [Read error: Connection reset by peer]
vivekp has joined #mlpack
dmatt has joined #mlpack
s1998_ has joined #mlpack
< s1998_> zoq: rcurtin: w.r.t PR 9 of models, I have changed the dataset to mnist (currently in csv but the train has size of 104MB, limit : 100MB). Should I read the data from original MNIST dataset (which was in bytes format) ?
< s1998_> Or should I break the train data (in csv) into two parts and then push the changes ?
< s1998_> Another thing is the current implementation reaches test accuracy of 82% but I think this can be fixed using (batch) normalization (since currently only l2 normalization is used). Should I do this (as in write code to find mean and sigma) or use batch norm layer ?
dmatt_ has joined #mlpack
dmatt has quit [Ping timeout: 255 seconds]
s1998_ has quit [Ping timeout: 260 seconds]
s1998_ has joined #mlpack
dmatt_ has quit [Remote host closed the connection]
daivik has joined #mlpack
Atharva has quit [Quit: Connection closed for inactivity]
daivik has quit [Quit: http://www.kiwiirc.com/ - A hand crafted IRC client]
daivik has joined #mlpack
daivik has quit [Quit: http://www.kiwiirc.com/ - A hand crafted IRC client]
< zoq> Atharva: I use almost the same setup irssi + tmux.
< zoq> s1998_: Testing the byte format sounds reasonable to me, hdf5 might be another solution.
< zoq> s1998_: About the accuracy, I think there are multiple options we could test, different architecture, l2 norm, batchnorm, another optimizer, so if you like please feel free to look into some options.
s1998_ has quit [Ping timeout: 260 seconds]
witness has joined #mlpack
ImQ009 has quit [Quit: Leaving]
< manthan> rcurtin : when exactly is the gradient() called for a differentiable layer?
< zoq> manthan: After the backward step.
< manthan> i mean what exactly will be the difference in the backward and gradient for a layer?
< manthan> backward will contain the update rule for backward pass
< manthan> what will gradient contain exactly?
< zoq> The update step for the parameter, you could merge both steps into one, but in this case you would have to run the backward step (error calculation) for the first layer as well, which is unnecessary since the error isn't going to be used.
daivik has joined #mlpack
< manthan> so the error obtained in this function is the error upto the present layer and we have to write the logic for updating the parameter given the error and input?
< zoq> correct
< rcurtin> zoq: would it be right to say that Backward() is the derivative of the inputs with respect to the error, whereas Gradient() is the derivative of the parameters with respect to the error?
< rcurtin> or to be clear, "Backward() is the derivative of the inputs of a particular layer with respect to the backpropagated error"
< zoq> yes, could could say that
< rcurtin> ok, just making sure---when I realized that it made the whole system a lot more clear to me, but I wasn't sure if I was correct :)
< zoq> Might be a good idea, to clarify that in the tutorial.
< zoq> Will set that on the list.
< manthan> shouldnt backward() be derivative of the backpropagated error with respect to the present layer paramters?
< manthan> so that i can backpropogate this error to previous layer
< manthan> for eg :- for ith layer, w(i)(new) = w(i)(previous) + alpha*dL/dw(i) and backpropogated error to previous layer will be dL/dw(i+1) * dw(i+1)/dwi ?
< manthan> so backward() implements logic for finding dw(i+1)/dwi given dL/dw(i+1)
< zoq> of a particular layer, in case of a ffn it's the previous one
< manthan> is this correct?^
< zoq> yes, looks correct to me
daivik has quit [Quit: http://www.kiwiirc.com/ - A hand crafted IRC client]
< zoq> haven't checked the backward/gradient step of the flexible ReLU layer.
< zoq> Will take a closer look at the code in the next days.
daivik has joined #mlpack
< rcurtin> the backward step looked correct to me when I did the previous review, but I am not 100% certain, only about 95% :)
< manthan> i think backward is correct but i am not sure about gradient
< manthan> this is because i am not able to clearly understand what gradient should contain
< manthan> backward is clear to me now^
< rcurtin> I think the gradient here should contain just one element, d L / d alpha
< manthan> yes it contains one element but what is error in this case which the function obtains as a function argument?
< manthan> i mean gradient() function^
< manthan> backward() - derivative of backprop error with respect to the input and gradient() - derivative of error with respect to the trainable parameter and what i wrote for backward above should be true for gradient()
< manthan> this is what the definitions look like from the various trainable layers that i saw
< manthan> with this, gradient() of flexible relu layer should always be 1 as flexible relu is (max(0,x) +a)
daivik has quit [Quit: http://www.kiwiirc.com/ - A hand crafted IRC client]
< manthan> rcurtin : zoq : i have updated the flexible relu gradient function now, pls have a look. the concept is clear to me now. Thanks.
< manthan> i think addition of this in the tutorial will be very useful for contributors :D
witness has quit [Quit: Connection closed for inactivity]
manthan has quit [Ping timeout: 260 seconds]