verne.freenode.net changed the topic of #mlpack to: http://www.mlpack.org/ -- We don't respond instantly... but we will respond. Give it a few minutes. Or hours. -- Channel logs: http://www.mlpack.org/irc/
Guest63658 has quit [Ping timeout: 265 seconds]
Guest63658 has joined #mlpack
prakhar_code[m] has quit [Ping timeout: 240 seconds]
killer_bee[m] has quit [Ping timeout: 245 seconds]
prakhar_code[m] has joined #mlpack
killer_bee[m] has joined #mlpack
sumedhghaisas has quit [Ping timeout: 255 seconds]
vivekp has joined #mlpack
< ShikharJ> rcurtin: Actually I'm trying to plot the GAN output to see for myself.
< Atharva> I figured it out. Can somebody confirm if this is correct. The delta matrix for linear layer is actually (weight.T * error) % derivative of the activation function, but in mlpack's implementation of anns, the activation functions are different layer objects so the delta matrix just becomes weight.T * error.
witness_ has joined #mlpack
govg has quit [Ping timeout: 256 seconds]
govg has joined #mlpack
< rcurtin> ShikharJ: ah, ok, in that case the preprocess_split way may not be the best way to go :)
< rcurtin> Atharva: I think that is correct, but I am not 100% sure, maybe zoq can confirm
witness_ has quit [Quit: Connection closed for inactivity]
sumedhghaisas has joined #mlpack
sumedhghaisas2 has joined #mlpack
sumedhghaisas has quit [Ping timeout: 268 seconds]
vivekp has quit [Read error: Connection reset by peer]
manish7294 has joined #mlpack
vivekp has joined #mlpack
manish7294_ has joined #mlpack
< manish7294_> rcurtin: Thanks, Hopefully updating gradient totally solved the problem. Here is a result - https://pasteboard.co/HnB3o5x.png
manish7294 has quit [Ping timeout: 260 seconds]
< ShikharJ> lozhnikov: You there?
manish7294_ has quit [Ping timeout: 260 seconds]
sumedhghaisas2 has quit [Ping timeout: 240 seconds]
sumedhghaisas has joined #mlpack
< rcurtin> manish7294: looks good, that is with SGD?
manish7294 has joined #mlpack
< manish7294> rcurtin: yes :)
sumedhghaisas2 has joined #mlpack
sumedhghaisas has quit [Ping timeout: 240 seconds]
< rcurtin> great---I guess you are trying now with some larger datasets? if those work, I think maybe we should get some basic benchmarking times, then we can see how we can accelerate the algorithm
< rcurtin> I have some ideas for avoiding the impostor recalculation
< rcurtin> I will have to write them down and think about it though
< ShikharJ> zoq: You there?
manish7294 has quit [Ping timeout: 260 seconds]
sumedhghaisas2 has quit [Ping timeout: 256 seconds]
manish7294 has joined #mlpack
< manish7294> rcurtin: It would be great if we can reduce the cost of impostors recalculation.
< rcurtin> right, so there are a couple of approaches that we could use together
< rcurtin> the first is that, if we know the distance to the k+1'th impostor, we can place a bound on how much closer that impostor can get each iteration
< rcurtin> I haven't derived the bound, but we can say that if the matrix did not change too much, the impostors will all be the same, so there is no need to recalculate
< manish7294> rcurtin: Currently I have also tried with iris. rcurtin And final objective seems preety good
< rcurtin> another acceleration possibility is to only recalculate impostors for those points in the dataset where the impostors could have changed
< rcurtin> those two ideas could probably be combined
< rcurtin> a third possibility, which is an approximation, is to only recalculate impostors every N iterations of the optimization for some N
< rcurtin> there are lots of possible ideas, so I am not too worried about being able to get some speedup in the end
< rcurtin> just keep in mind, if you are thinking about MNIST, that nearest neighbor search is going to be slow for that dataset almost no matter what because it is so high dimensional
manish7294 has quit [Ping timeout: 260 seconds]
manish7294 has joined #mlpack
< manish7294> rcurtin: Everything sounds good :)
sumedhghaisas has joined #mlpack
manish7294 has quit [Ping timeout: 260 seconds]
< zoq> ShikharJ: Yeah.
< zoq> Atharva: That is correct.
< ShikharJ> zoq: I was wondering why in the GAN implementation we're just training the Generator on a single noise input (columns = 1), and not batch-wise (columns = batchSize)?
travis-ci has joined #mlpack
< travis-ci> manish7294/mlpack#12 (lmnn - 70680c7 : Manish): The build was broken.
travis-ci has left #mlpack []
< zoq> ShikharJ: I guess, since Kris created the PR, some things changed, like batch support for the conv layer, I think you worked on that part, so I agree adding batch support is something we should add.
< Atharva> zoq: Thanks for the confirmation.
sumedhghaisas2 has joined #mlpack
sumedhghaisas has quit [Ping timeout: 240 seconds]
< zoq> ShikharJ: If you need a system to run the code on for hours and hours, we could perhaps use one of the benchmark systems.
< ShikharJ> zoq: I think I found the reason behind that strategy here (https://github.com/mlpack/mlpack/pull/1066#issuecomment-322114951).
< ShikharJ> zoq: I didn't specifically work on batch support for CNNs, they just take a single input at one time, since we need them to take an input in a 2d matrix form and not as individual columns. I believe my concern is superficial here because of the pipeline that lozhnikov has created.
< ShikharJ> I'm guessing batch support was already there at Kris' time, since the discriminator does take batch based inputs, though it faces an error that I'm trying to fix.
< rcurtin> ShikharJ: I am not sure if this is a helpful comment that addresses what you are talking about, but I believe that batch support was added to the ANN framework after Kris's project, with the merge of #1137, which was in October 2017
< rcurtin> however, if I remember right, that was mostly a change to the optimizers themselves, not to the ANN framework... maybe there were minor changes there
< ShikharJ> rcurtin: I'll dig into this, thanks!
sumedhghaisas2 has quit [Ping timeout: 240 seconds]
< rcurtin> I'm not sure how useful looking through #1137 is, mostly I just wanted to point out that at the time of Kris's code, it would have been reasonable if he implemented it in such a way that he was only considering batches of size one
< rcurtin> but if it is helpful I am glad to have shared it :)
< zoq> If it works for batch size = 1 for now, thats fine we can work on this part later; if you like I can implement that part
< ShikharJ> rcurtin: It did help :)
< rcurtin> :)
< ShikharJ> zoq: Sure, we just need to check for code correctness for now and we can worry about the batch sizes for later. However, it does seem to be an interesting problem to solve :)
< zoq> agreed, I'll take another look into gradient step later today.
< ShikharJ> zoq: Also with lozhnikov's approach of a single noise input, the generator network doesn't need to have batch normalization layer, since the input is singular, so that means lesser computation.
< ShikharJ> zoq: Let's just hope we can find some good results on parameters. What about the benchmark systems?
< zoq> right, which is good for testing
< zoq> if you need a system to run the code?
< ShikharJ> Yes, are the online computing instances?
< ShikharJ> I have no clue regarding what benchmark systems are in mlpack?
< ShikharJ> zoq: Could you tell me more?
sumedhghaisas has joined #mlpack
< zoq> rcurtin: can we use one of the benchmark systems?
< rcurtin> sure, I think only I am allowed to have root on them because they are Symantec owned, but I can definitely create an account on one of them
< rcurtin> ShikharJ: basically, Symantec provides some number of build systems for us, and we have 5 systems that we use to benchmark mlpack through the benchmarks system: https://github.com/mlpack/benchmarks/
< rcurtin> but these systems are useful also for long-running jobs that might happen during GSoC
< rcurtin> zoq: how about savannah.mlpack.org?
< rcurtin> ShikharJ: let me know what username you like, then I'll get the account set up and PM you the credentials
< ShikharJ> rcurtin: Amazing! ShikharJ would be a good username :P Thanks for the help!
ImQ009 has joined #mlpack
sumedhghaisas has quit [Ping timeout: 240 seconds]
sumedhghaisas has joined #mlpack
sumedhghaisas2 has joined #mlpack
sumedhghaisas has quit [Ping timeout: 260 seconds]
< zoq> yeah, savannah works great
sumedhghaisas2 has quit [Ping timeout: 276 seconds]
sumedhghaisas has joined #mlpack
sumedhghaisas2 has joined #mlpack
sumedhghaisas has quit [Ping timeout: 255 seconds]
sumedhghaisas2 has quit [Ping timeout: 240 seconds]
sumedhghaisas has joined #mlpack
ImQ009 has quit [Read error: Connection reset by peer]
ImQ009 has joined #mlpack
ImQ009 has quit [Client Quit]
witness_ has joined #mlpack
travis-ci has joined #mlpack
< travis-ci> ShikharJ/mlpack#166 (GAN - 01316ea : Shikhar Jaiswal): The build has errored.
travis-ci has left #mlpack []
sumedhghaisas2 has joined #mlpack
sumedhghaisas has quit [Ping timeout: 244 seconds]
sumedhghaisas has joined #mlpack
sumedhghaisas2 has quit [Ping timeout: 256 seconds]
sumedhghaisas2 has joined #mlpack
sumedhghaisas has quit [Ping timeout: 240 seconds]
sumedhghaisas has joined #mlpack
sumedhghaisas2 has quit [Ping timeout: 240 seconds]
sumedhghaisas has quit [Ping timeout: 248 seconds]
sumedhghaisas has joined #mlpack
sumedhghaisas has quit [Ping timeout: 244 seconds]
sumedhghaisas has joined #mlpack
sumedhghaisas has quit [Read error: Connection reset by peer]
sumedhghaisas has joined #mlpack