verne.freenode.net changed the topic of #mlpack to: http://www.mlpack.org/ -- We don't respond instantly... but we will respond. Give it a few minutes. Or hours. -- Channel logs: http://www.mlpack.org/irc/
witness_ has quit [Quit: Connection closed for inactivity]
vivekp has joined #mlpack
witness_ has joined #mlpack
< ShikharJ> zoq: Are you there?
< jenkins-mlpack> Yippee, build fixed!
< jenkins-mlpack> Project docker mlpack nightly build build #366: FIXED in 2 hr 57 min: http://masterblaster.mlpack.org/job/docker%20mlpack%20nightly%20build/366/
< zoq> ShikharJ: Great results, will take a look at the PR later today.
witness_ has quit [Quit: Connection closed for inactivity]
< ShikharJ> zoq: I was hoping if we could discuss the design for the DualOptimizer PR? If you'll be free later during the day, please ping me then.
< zoq> ShikharJ: If you like we can talk about the design now.
< Atharva> zoq: For VAE models, MNIST data is needed, should we store it as csv in the models repo or should we store it as the original idx-ubyte format
< Atharva> Another option can be to download it the first time someone builds the models repo
< ShikharJ> Atharva: I have the tranposed MNIST data available in csv format, let me know if you need it.
< Atharva> ShikharJ: What do you mean by transposes here?
< Atharva> transposed*
< ShikharJ> Every image is a column of size 784. The original dataset had images laid across rows.
< zoq> Atharva: csv is fine hdf5 might be another option, we could compress the dataset and uncompress it as a build step, something like: https://github.com/mlpack/mlpack/blob/6fd5e527b54bf83993f98f8ea894734aa620bb62/src/mlpack/tests/CMakeLists.txt#L185-L189
< ShikharJ> So the dataset has dimensions 70,000 x 784 instead of 784 x 70,000.
< Atharva> ShikharJ: got it :), I thought by transposed you meant something else in this case
< Atharva> zoq: Thanks, can armadillo load hdf5 files?
< zoq> Atharva: Yes, build you have to build armadillo with hdf5 support
< Atharva> zoq: Okay, If hdf5 doesn't have a size much lesser tan csv, I guess it's better to go with csv as otherwise people will have to build armadillo differerntly for this one task
< zoq> Atharva: Agreed.
< Atharva> ShikharJ: Where do you have it? on some repo?
< ShikharJ> Atharva: On my laptop, and on savannah server.
< Atharva> Can you give me the link?
< ShikharJ> Atharva: If there's a server you need it on, I can scp the zip file?
< ShikharJ> Atharva: The zip is about 17 MBs, so maybe I can send it over mail as well
< zoq> ShikharJ: We can put it in the jenkins folder.
< ShikharJ> zoq: I'm not sure if I understand what you mean by jenkins folder?
< ShikharJ> Did you mean jenkins-conf repository?
< zoq> If we move the file in the jenkins workspace, somene can download it over http.
< zoq> This is just another possibility.
< ShikharJ> zoq: Ideally we should try and upload the dataset directly to mlpack/models?
< zoq> ShikharJ: Agreed, what's the size of the compressed dataset (tar.gz)?
< ShikharJ> zoq: I only have the zip file with me, and that's about 17 MBs.
< ShikharJ> I'm not sure if tar ball would have a significantly different size or not.
< zoq> ShikharJ: Right, it's just that on some systems you have to install unzip or something like that to extract the archive.
< ShikharJ> zoq: I'll upload the tarball to mlpack/models. Atharva, you can download from there then.
< zoq> ShikharJ: Great, thanks!
< Atharva> ShikharJ: Thanks!
< ShikharJ> Interesting, the tarball is about 13.9 MBs in size.
< Atharva> How much had you thought it should be?
< ShikharJ> Atharva: I wasn't expecting it to be about 5/6th the size of a zip file.
< Atharva> Ohh, okay
< ShikharJ> zoq: I have opened a pull request for the dataset. Atharva, you can download from the same once it is merged.
< ShikharJ> zoq: Are you still around, maybe we can discuss the design of the dual optimizer further?
< zoq> ShikharJ: Sure, I'm here.
< ShikharJ> zoq: As far as I can see, when we create an optimizer object, it internally calls the optimize function. But what that optimize function does is not clear to me.
< ShikharJ> zoq: More specifically, what functionality does optimize() offer?
< zoq> ShikharJ: The optimizer will call the Evaluate function to get the current loss, in our case this will run the Forward pass. Afterwards the optimizer will call the Gradient function to get the gradients for the update step, so this will call the Forward/Backward and Gradient function in case of a network.
< zoq> See the example optimizer.
< zoq> In our case ObjectiveFunction is the GAN class.
< ShikharJ> zoq: Alright, so if we create two optimizers, we'll need to provide two evaluate, two gradient, and two forward functions (because we're calling discriminator and generator forward at the same time inside GAN::Forward) right/
< ShikharJ> ?
< zoq> That's an option, but we could use the same Evalaute function if we could distinguish between the functions somehow.
< zoq> ShikharJ: If you like I can write a dummy class with the idea I have in mind.
< ShikharJ> zoq: Hmm, I can't think of an idea that would be fast regarding this, do you have an idea how can this be achieved. I can only think of a template based solution, but that woud lead to runtime loss.
< ShikharJ> zoq: Please go ahead, I can't think of a good solution for this problem.
< zoq> Sure, maybe I missed something :) I'll see if I can put some time into the idea later today.
< ShikharJ> zoq: Also regarding the models directory, what name would you prefer? Maybe a generic name like datasets/ ?
< zoq> yeah, or data
< ShikharJ> I'll make the changes and push again.
< zoq> ShikharJ: Okay, thanks again.
petris has quit [Remote host closed the connection]
petris has joined #mlpack
petris has quit [Remote host closed the connection]
petris has joined #mlpack
witness_ has joined #mlpack
< ShikharJ> zoq: I was thinking, in the meantime, we can try completing the RBM PR?
< zoq> ShikharJ: I think that is a great idea :)
< ShikharJ> zoq: Great, I have updated the WGAN PR, and I'll update the blog with a post as well.
< zoq> ShikharJ: Awesome, I'll take a second look at the changes later today; and we should be able to merge the code in the next days.
< Atharva> ShikharJ: The dataset you uploaded doesn't have the labels, we don't need it for generative models but I guess it will be better to have them.
< Atharva> Or do you need it for GANs, sorry I am not sure
< Atharva> ?
< ShikharJ> Atharva: Yeah, I did that to prevent having problems for the GAN code. We can provide them separately, I have the original full dataset as well.
< Atharva> Oh cool, maybe you could commit the csv with labels to the PR :)
< Atharva> No rush, do it when you get time
< ShikharJ> Atharva: For most other cases, the kaggle_train_test_dataset.zip should suffice, I don't really see a need for adding labels, as it has them I think.
< Atharva> Ahh yes, it has. No problem then.
vivekp has quit [Ping timeout: 245 seconds]
ImQ009 has joined #mlpack
petris has quit [Quit: Bye bye.]
petris has joined #mlpack
ImQ009 has quit [Quit: Leaving]
< Atharva> sumedhghaisas: You there?