verne.freenode.net changed the topic of #mlpack to: http://www.mlpack.org/ -- We don't respond instantly... but we will respond. Give it a few minutes. Or hours. -- Channel logs: http://www.mlpack.org/irc/
Samir has joined #mlpack
< Samir> Hello, I want to contribute to mlpack during my summer time and learn more about reinforcement learning, Is there anyone currently working on the reinforcement project (which was an idea of GSoC 18), I am interested in getting experience and more knowledge
< Samir> I am currently building the repository
Samir|Phone has joined #mlpack
Samir has quit [Killed (cherryh.freenode.net (Nickname regained by services))]
Samir has joined #mlpack
Samir has quit [Quit: Leaving]
Samir has joined #mlpack
Samir|Phone has quit [Quit: Mutter: www.mutterirc.com]
Samir_ has joined #mlpack
Samir has quit [Read error: Connection reset by peer]
Samir_ is now known as Samir
yaswagner has quit [Ping timeout: 252 seconds]
< Atharva> zoq: I trained it using RMSProp and the results are much better,
< Atharva> but I trained it much more I think, when I used Adam, I stopped it after about 3-4 hours. Last night, I trained it for about 7.
< Atharva> To make sure, I will train it tonight again using Adam for 7 hours and see the results.
< zoq> Atharva: Thanks for the update, do you have any numbers?
< Atharva> zoq: Yes, With RMSProp, after about 120 epochs, loss went down to ~107 (which includes kl divergence). With Adam, after 50 epochs, it went down to ~120(includes kl divergence)
< zoq> Atharva: Okay, thanks!
< Atharva> zoq: Tonight, I will train with Adam for more than 100 epochs and see how much the loss goes down.
< zoq> Atharva: Good, if need a machine to run some experiments, let us know.
< Atharva> zoq: Thanks! Maybe I will need it when I train on celebA dataset. Right now, my machine suffices :)
< sumedhghaisas> Atharva: Hi atharva
< sumedhghaisas> Was just looking at the PR. Did you decide the convolutional architecture or took it from some paper?
< sumedhghaisas> Also could we open a seperate PR for Bernoulli? That way the current PR of ReconstructionLoss can be just put on hold while we merge the new PR
< sumedhghaisas> The current ReconstructionLoss PR is I think fully reviewed, its better to just hold it for now
< Atharva> sumedhghaisas: Yeah, I will open a seperate PR.
< Atharva> About the convolutional architecture, I just saw it on some implementation on github, that too I didn't copy completely. It's just the first experiment so I didn't focus much on the architecture.
< Atharva> Do you have some suggestions about how I should better decide the architecture?
< sumedhghaisas> Atharva: The architecture is fine. :)
< sumedhghaisas> there are couple of good one which are used for state of the art on BinaryMnist but for Mnist this will do
< sumedhghaisas> its just that, we should mention here if we have taken the architecture from somewhere.
< Atharva> I haven't really taken it from anywhere, so I guess it's ok
< sumedhghaisas> Also amazing work on models, we finally did it.I just have couple of small comments on the PR.
< sumedhghaisas> Atharva: Then its fine. :)
< Atharva> Thanks for the all the help! Couldn't have been possible without it :)
< sumedhghaisas> Binary Mnist is more exciting, we can actually compare the results with state of the art and also merge in ReconstructionLoss
< Samir> Hello guys, can I work on this project (https://github.com/mlpack/mlpack/wiki/SummerOfCodeIdeas#reinforcement-learning)? is it available or there is someone working on it?
< sumedhghaisas> I think the PR looks good, just have documentational changes from my side.
< Atharva> Yeah, I tried training a model on binary mnist with reconstructionLoss, and the loss goes down till about ~120 out of which ~10 is kl
< Atharva> Haven't generated any images though, I will do that sson
< Atharva> soon
< sumedhghaisas> great...
< sumedhghaisas> this is feed forward network?
< Atharva> Yeah, not conv
< sumedhghaisas> hmm... should go till 100 though.
< sumedhghaisas> what is the network?
< sumedhghaisas> Samir: Hi Samir.
< Atharva> The same network in vae.cpp
< Atharva> I didn't train it long enough
< Samir> Hi sumedhghaisas :)
< Atharva> Will do it again tonight
< sumedhghaisas> Samir: Marcus or @zoq is handling the specific project as a mentor. You can ask him for any specific questions. :)
< sumedhghaisas> I don't have much information about it, sorry. :(
< sumedhghaisas> can I see the binarization code that you used?
< sumedhghaisas> Atharva:
< Samir> sumedhghaisas, Thanks :), I want to contribute to mlpack during my summer time, I will ask them when they are around.
< sumedhghaisas> Samir: Happy to help in any way I can. :) Let me know if you have any other questions regarding the constribution.
< Atharva> sumedhghaisas: Do you mean the code I used to prepare binary MNIST from normal MNIST?
< Atharva> Another thing, using the convolutional model, although the total loss goes down, the kl loss is more than in feedforward models
< Atharva> Last night, I trained using RMSProp and the error went down to 107, the results look much better than the conv model
< Atharva> I think it's because the distribution is not fitting the standard normal well
< Atharva> So, sampling from the prior doesn't give very good results
< sumedhghaisas> KL discrepancy is expected.
< sumedhghaisas> Got a meeting in 2 mins.
< sumedhghaisas> Can I get back to you in 2 hours?
< Atharva> It’s okay. Actually, I will be a little busy tonight
< Atharva> Maybe we will talk tomorrow
ImQ009 has joined #mlpack
< zoq> Samir: Hello, would be great to push the RL project, so if you are interested let me know what you have in mind and we can go from there.
killer_bee[m] has left #mlpack []
< sumedhghaisas> Atharva: Sure
travis-ci has joined #mlpack
< travis-ci> mlpack/mlpack#5343 (master - 455b009 : Ryan Curtin): The build was fixed.
travis-ci has left #mlpack []
ImQ009 has quit [Quit: Leaving]