verne.freenode.net changed the topic of #mlpack to: http://www.mlpack.org/ -- We don't respond instantly... but we will respond. Give it a few minutes. Or hours. -- Channel logs: http://www.mlpack.org/irc/
< kris__> lozhnikov: Do you have the script that you used for generating the images and from the rbm with deep learning.net example. Also the cpp file of the test. I seem to have it missing.
kris1 has quit [Quit: kris1]
kris1 has joined #mlpack
kris1 has quit [Quit: kris1]
< lozhnikov> kris__: Try the following file. https://www.irccloud.com/pastebin/yKmWeiQu/main.cpp
kris1 has joined #mlpack
< lozhnikov> kris__: I wrote that for the ssRBM but it isn't difficult to use the binary RBM instead.
< lozhnikov> Could you describe your build issue more detailed?
< kris__> The build issue had to do with the cstdin not found since i think pywrapper were introduced that require that library.
< kris__> I was able to build after removing the line from CmakeLists.txt that attaches -stdc++=libc++ when the apple mac os version is less that 10.9. I remove the if statement and it worked.
< lozhnikov> which branch are you compiling?
< kris__> The newest branch. Since then i have reverted the changes to the merge i did with latest version. now i run 2.2.4 everything works fine now.
< lozhnikov> I haven't got MacOS. So, probably it is difficult to reproduce
< kris__> Okay no problem. Right now i got working using the easy solution. Will look at a detailed solution later.
< kris__> Both the resize layer and cross entropy with logit are succesfully build. Travis is faililng on randomForestTest.
< kris__> So you can go ahead and have a look at them.
< lozhnikov> I'll look through that today
< kris__> ok, i will give the orilley example a final try today....
< kris__> The orilley example is working now check it out here https://gist.github.com/kris-singh/4b355418edd9c69ede11c4af18086438
< kris__> since, CrossEntropyWithLogits with sigmoid is same as CrossEntropy so i replaced it in the generator network.
< kris__> ./gan_test2.o -i train7.txt -o output.txt -e 200 -m 200 -x 1 -N 100 -r 0.003 -v
< kris__> If you can run it with ./gan_test2.o -i train7.txt -o output.txt -e 20 -m 2000 -x 1 -N 100 -r 0.003 -v these parameters that would be great.
< lozhnikov> okay, I'll run the test soon
< lozhnikov> kris__: I ran the test with the following arguments: -e 20 -m 2000 -x 300 -N 100 -r 0.0003 -v
< lozhnikov> I think these parameters correspond to the oreilly example better than your parameters
< kris__> Sure, but i think these would require a lot time to converge i was going for fining a good set of starting parameters and optimising them.
< kris__> Though i think e 20 is okay but i think x 300 is pretty high.
< lozhnikov> I think we haven't got enough time for obtaining new parameters
< lozhnikov> okay, I'll replace that by 100
kris1 has quit [Quit: kris1]
kris1 has joined #mlpack
kris1 has quit [Quit: kris1]
kris1 has joined #mlpack
kris1 has quit [Client Quit]
kris1 has joined #mlpack
< lozhnikov> [INFO ] gradientGenerator = 0.000000e+00
< lozhnikov> [INFO ] gradientDiscriminator = 2.155897e-02
< lozhnikov> [INFO ] gradientGenerator = 0.000000e+00
< lozhnikov> [INFO ] gradientDiscriminator = 2.877930e-02
< lozhnikov> kris__: the gradient of the generator is zero
< lozhnikov> I tried to replace GaussianInitialization gaussian(0, 1); by GaussianInitialization gaussian(0, 0.02);. Actually, the oreilly example uses 0.02 as the standard deviation
< lozhnikov> I got the same result
< lozhnikov> Curiously, SigmoidLayer + CrossEntropyError works fine with it
< lozhnikov> so, the error that we observed yesterday happened due to incorrect deviation
kris1 has quit [Quit: kris1]
< lozhnikov> however, CrossEntropyErrorLogits works with deviation 1.0
kris1 has joined #mlpack
kris1 has quit [Client Quit]
kris1 has joined #mlpack
< kris__> Just saw the comments...
< kris__> So does the sigmoid + cross entropy layer work.....
< lozhnikov> sigmoid + cross entropy shows the same results as CrossEntropyErrorLogits with deviation 0.02
< lozhnikov> but the gradient of the generator is zero
< kris__> Okay i will check what wrong now.
< kris__> btw i have updated the summary of the work. Will you have a look at it when you have time.
< lozhnikov> actually, that isn't the summary
< lozhnikov> you should write the final blog post and describe all changes you have done
< lozhnikov> end then you should submit the final evaluation with the link to the final blog post
< kris__> I do not know why it doesn't show up on the website though.
< lozhnikov> try to change "Summary Date" by "Date"
< lozhnikov> 1. "I am happy to say that in terms of visual reconstruction in both the examples of Mnist digit generation and Gaussian distribution generation we were able to get comparable results with keras and tensorflow."
< lozhnikov> I think we didn't get good results on the mnist dataset yet.
< lozhnikov> 2. "Here are the results of our implementation on the digits dataset(smaller version Mnist)."
< lozhnikov> Actually, this is the mnist dataset
< kris__> yes mnist 7 i would update it. Okay i will try to run it for more number of epochs and see if we can get better results.
< lozhnikov> 3. "One of the reasons accuracy of ssRBM is less than ssRBM".
< lozhnikov> I guess "is less than binary RBM" is correct
kris1 has quit [Quit: kris1]
< lozhnikov> 4. "I would also like to say here that Mikhail tried to convert mu-ssRBM code for testing our implementation but it took a lot of time finally."
< lozhnikov> I finished the implementation of the mu-ssRBM. But it isn't well tested yet, I focused on the GAN PR. That's why I haven't pushed the code yet. So, I think there is no sense to write about that. However, it's up to you.
< lozhnikov> 5. "We tried ssRBM on the cifar data set code but due to the large volume of data set and scarcity of the computation resources, we decided that it was not really required."
< lozhnikov> If I remember right, last time you told that you got good accuracy on the CIFAR dataset.
< zoq> kris__: About the libc++ issue, what OS did you use?
< zoq> kris__: Also the static site generator is really picky about the metadata header, if you remove the extra lines (between the meta data) the build should pass
< lozhnikov> kris__: 6. I think it is reasonable to upload the tests to github gists and mention them in the blog post.
< lozhnikov> Except the comments above the blog post looks good to me
< lozhnikov> kris__: I looked through the oreilly example again. The fourth ReLU layer is not needed in the generator network
klitzy has joined #mlpack
klitzy has quit [Client Quit]
< kris__> I will have to look at the cifar classification accuracy
< kris__> I have the code for it
< kris__> But I Have to look at the results once
< kris__> Zoq I use Mac OS X 10.12 if I rember right
< lozhnikov> kris__: without the fourth ReLU layer the gradient of the generator network isn't equal to zero
< lozhnikov> [INFO ] gradientGenerator = 3.100361e+00
< lozhnikov> [INFO ] gradientDiscriminator = 1.375701e-04
< lozhnikov> [INFO ] gradientGenerator = 3.084291e+00
< lozhnikov> [INFO ] gradientDiscriminator = 1.281130e-04
< lozhnikov> [INFO ] gradientGenerator = 3.063628e+00
kris1 has joined #mlpack
< kris__> yup 10.12.6build version 16G29
< zoq> kris__: okay, let me update my machine and see if I can reproduce the issue
< kris__> zoq: can we use l1 or l2 weight regularisers in the ffn.
< zoq> kris__: Unfortunately it's not implemented yet.
kris1 has quit [Quit: kris1]
kris1 has joined #mlpack
kris1 has quit [Quit: kris1]
kris1 has joined #mlpack
< lozhnikov> kris__: I ran the test with the following arguments -e 20 -m 2000 -x 100 -N 100 -r 0.03 -v and got the following results https://usercontent.irccloud-cdn.com/file/fuO7OT6K/mnist-conv.png
< lozhnikov> something went wrong
< kris__> Well i don't know. I think orilley example goes through around 6000 epoch's to the desired result.
< kris__> How did your results converge so fast.
< kris__> I am running a much smaller epoch but it is being going on since morning.
< lozhnikov> I think something is incorrect since the gradients differ too much
< lozhnikov> [INFO ] gradientGenerator = 1.562875e+00
< lozhnikov> [INFO ] gradientDiscriminator = 8.360856e-06
< lozhnikov> why the objective is so huge?
< lozhnikov> did you fix the standard deviation?
< kris__> No i was training on the CrossEntropy with Logits.
< kris__> since the training was going on since morning i din't kill it.
< kris__> I will kill it now.
< lozhnikov> how much time should the test take?
< kris__> I am not sure. I actually did not run the orilley example since you said that you had run it already.
< lozhnikov> Mini-batch SGD provides an output each epoch. So, you can estimate the time
< kris__> Okay i will check and tell you but i am pretty sure that one epoch takes > 25 min on my machine with m = 2000.
< kris__> lozhnikov I had a question you said that mnist 7 digit reconstruction we got was not good. I think that could be because of the regularisation of weights.
< lozhnikov> kris__: and what is the question?
< kris__> No i was just saying that we don't have regularisation of weights in mlpack so i was trying with higher number of epoch to get a images what were smooth.
< kris__> But i am unable to do that.
< lozhnikov> maybe you are right. I didn't dig into the keras example
< lozhnikov> kris__: have you fixed the blog post and have you submitted the final evaluation? I just want to make sure that you haven't got troubles with that
< kris__> Not yet. I will do that by tonight. I just wanted to wait for the results from the cnn gan but i think that would not be possible i would go ahead.
< kris__> Should i submit it now. Or wait till tuesday. Because if you submit once you can't change it.
< lozhnikov> I don't recommend to wait Tuesday
< lozhnikov> so, I think it is better to complete the evaluation now
< kris__> Ok i will do that in hour or so then.
vpal has joined #mlpack
vpal has quit [Read error: Connection reset by peer]
vivekp has quit [Ping timeout: 248 seconds]
vivekp has joined #mlpack
vivekp has quit [Ping timeout: 260 seconds]
vivekp has joined #mlpack
< kris__> zoq: The images are not displayed in the mlpack.org page.
< kris__> but on github page they are shown do i need to give full path for img src right now i have given a relative path only.
vivekp has quit [Read error: Connection reset by peer]
< lozhnikov> try to fix the path i.e. replace "../images/mnist_out.png" by "images/mnist_out.png"
< lozhnikov> take a look at the fourth blog post https://github.com/mlpack/blog/blob/master/content/blog/KrisWeekFour.md
vivekp has joined #mlpack
< kris__> Done...thanks.
< kris__> lozhnikov: Should i implement the batch norm layer for the cnn test. Also do you have any ideas how we could test it faster.
< kris__> Also one more thing was batch normalization acts as a regularizer it can be replaced by the DropOut layer. But when i was using the dropout layer the network converged very quickly i mean in 2-3 iterations. This is true for even individual ffn's.
< kris__> zoq: Why do you say the present implmentation of batchNorma won't work with conv layers.
< kris__> *batchNorm
vivekp has quit [Ping timeout: 276 seconds]
< zoq> kris__: The current implementation does not cover the data dimensions correctly if > 2 (normalization step isn't correct); Sumedh is working on that end.
< zoq> kris__: Looks like you fixed the image issue?
< zoq> kris__: I'm not able to reproduce the libc++ issue, Mac OS > 9 (Mavericks) should automatically link against libc++, did you say that the problem encountered with mlpack 2.2.5?
< kris1> Yes the newest version.
< zoq> Did you build with -DBUILD_PYTHON_BINDINGS=OFF?
< kris1> no actually i did not…..
< zoq> Do you have time to test that?
< zoq> not sure that's the problem
< kris1> Well that would fix it i think
< kris1> I will check in the morning tommrow. I have a early day tommrow.
< zoq> okay, thanks a lot
vivekp has joined #mlpack
vivekp has quit [Ping timeout: 255 seconds]
vivekp has joined #mlpack
kris1 has quit [Quit: kris1]
kris__ has quit [Quit: Connection closed for inactivity]