ChanServ changed the topic of #mlpack to: "mlpack: a fast, flexible machine learning library :: We don't always respond instantly, but we will respond; please be patient :: Logs at http://www.mlpack.org/irc/
gotadachi has quit [Quit: Leaving...]
gotadachi has joined #mlpack
< jenkins-mlpack2> Yippee, build fixed!
< jenkins-mlpack2> Project mlpack - git commit test build #443: FIXED in 1 hr 37 min: http://ci.mlpack.org/job/mlpack%20-%20git%20commit%20test/443/
< KimSangYeon-DGU[> The training time took about an hour and a half
< kartikdutt18[m]> Hmm, I can't get even a single epoch in an hour. The model is bit different from darknet 19.
< KimSangYeon-DGU[> Yes, it's simpler model than darknet 19
< KimSangYeon-DGU[> Maybe, there is bottleneck in the darknet implementation.
< kartikdutt18[m]> What do you suggest?
< KimSangYeon-DGU[> When you trained, can you use multiple cores?
< KimSangYeon-DGU[> * When you trained, could you use multiple cores by any chance?
< kartikdutt18[m]> I am using open mp.
< KimSangYeon-DGU[> Ok
< kartikdutt18[m]> I'll just check the number of threads.
< KimSangYeon-DGU[> My laptop spec is i7, 16gb (8 cores)
< kartikdutt18[m]> I am using 12 threads according to activity manager.
< kartikdutt18[m]> Should I send you the training subset or use zoq's machine if they would be faster?
< KimSangYeon-DGU[> Let me try to train from the scratch. Can I reproduce the training work by using the darknet model PR?
< KimSangYeon-DGU[> Did you make any change in your local repo?
< KimSangYeon-DGU[> ,not remote repo
< kartikdutt18[m]> Let me just push it again ( I think only training params might be different).
< KimSangYeon-DGU[> Ok, and I found in darknet framework, the loading time is so fast. It only loaded images as batch size
< KimSangYeon-DGU[> The batch size was 128
< kartikdutt18[m]> We load the whole dataset first and then train the model.
< KimSangYeon-DGU[> Right
< KimSangYeon-DGU[> I'll reproduce the small network in mlpack and let's see what's happening
< kartikdutt18[m]> Sure, Let me know if I need to make the model.
< KimSangYeon-DGU[> :)
< kartikdutt18[m]> I can reuse the functions in darknet class.
< KimSangYeon-DGU[> But, your laptop is already running to train the model, so I think it's good to leave it as it is for the time being.
< kartikdutt18[m]> Sure :)
< KimSangYeon-DGU[> Without stop
< KimSangYeon-DGU[> :)
< KimSangYeon-DGU[> I was trying to find any Darknet 19 references except for the official one, but I couldn't anything yet. I'll share it with you, if I find.
< kartikdutt18[m]> Thanks, I'll check if we can load images in parallel as well. Will keep posting updates about training.
< KimSangYeon-DGU[> Thanks!
ImQ009 has joined #mlpack
< KimSangYeon-DGU[> kartikdutt18: Let me train the model using Darknet 19 in the Darknet framework first.
< kartikdutt18[m]> Ok, on Cifar 10?
< KimSangYeon-DGU[> Yes
< kartikdutt18[m]> I think you would have to change a few pooling layer (increase padding) because for 32 x 32 the size will go to 0.
< KimSangYeon-DGU[> Yes, I think so
< kartikdutt18[m]> This might be [useful](https://github.com/mlpack/models/pull/20#discussion_r439991844).
< KimSangYeon-DGU[> Great, thanks 👍️
< KimSangYeon-DGU[> kartikdutt18: Started, the Darknet 19 takes about 14 secs per 1 iteration.
< KimSangYeon-DGU[> It's much slower than the simple model that I've showed you
< KimSangYeon-DGU[> * kartikdutt18: Started, for your information, the Darknet 19 takes about 14 secs per 1 iteration.
< KimSangYeon-DGU[> * kartikdutt18: Started. For your information, in the Darknet framework the Darknet 19 takes about 14 secs per 1 iteration.
< KimSangYeon-DGU[> * kartikdutt18: Started. For your information, in the Darknet framework, the Darknet 19 takes about 14 secs per 1 iteration.
< kartikdutt18[m]> Hmm, It takes nearly 30-45 sec for 1 iteration. (1200 iteration in an epoch)
< kartikdutt18[m]> @Kim
< kartikdutt18[m]> KimSangYeon-DGU: , What's the initial loss?
< KimSangYeon-DGU[> 7.350986
< kartikdutt18[m]> I think that's also random initialization. I guess you also used linear layer to convert last 1000 channels to 10?
< KimSangYeon-DGU[> Oops, I didn't !
< KimSangYeon-DGU[> kartikdutt18: This is the summary
< KimSangYeon-DGU[> - https://pastebin.com/DUzP0fJv
< saksham189Gitter> @kartikdutt18 Did you find any examples of using Darknet on CIFAR dataset online? Do you think we should try with a different dataset?
< KimSangYeon-DGU[> And the initial loss is `2.419931`
< kartikdutt18[m]> I didn't find darknet 19 or 53 with cifar 10. There is only darknet small. I think I got nearly the same loss with Xavier Initialization but it didn't change after 2 epochs for various learning rates.
< KimSangYeon-DGU[> I'm training the model with cifar-10 and Darknet 19.
< kartikdutt18[m]> Other dataset that could be possible is [tiny imagenet](https://www.kaggle.com/c/tiny-imagenet/data).
< KimSangYeon-DGU[> kartikdutt18: Yes, but I guess Saksham wants the benchmark.
< kartikdutt18[m]> We do have some reference for darknet 19 on complete imagenet dataset. [Link](https://pjreddie.com/darknet/imagenet/)
< KimSangYeon-DGU[> saksham189 (Gitter): Would it be better for us to change the dataset to imagenet?
< saksham189Gitter> Yes but I guess the training time would be even longer, right?
< KimSangYeon-DGU[> I think so
< kartikdutt18[m]> I think so, yes
< saksham189Gitter> and Imagenet has 1000 classes so, definitely alot harder problem.
< kartikdutt18[m]> Yes.
sakshamb189[m] has joined #mlpack
< KimSangYeon-DGU[> I'm training the Darknet-19 model for CIFAR-10 in the Darknet framework. After the first epoch, let me check the validation error.
< KimSangYeon-DGU[> * I'm training the Darknet-19 model for CIFAR-10 in the Darknet framework. After the first epoch, let me check the validation accuracy.
< KimSangYeon-DGU[> To make the benchmark
< KimSangYeon-DGU[> Currently, the loss is not decreasing
< KimSangYeon-DGU[> * Currently, the loss is not decreasing easily
< kartikdutt18[m]> I think the graph would look similar to this. [Link](https://github.com/mlpack/models/pull/20#issuecomment-646528999)
< KimSangYeon-DGU[> Yes so far
< KimSangYeon-DGU[> We can decrease the channel size as the next experiment.
< KimSangYeon-DGU[> Not now :)
< KimSangYeon-DGU[> I guess CIFAR-10 image size 32 x 32 is so small for the current Darknet 19 architecture.
< KimSangYeon-DGU[> Hm, imagenet is so big... so I'm thinking to see if it's possible to migrate the pretrained Darknet-19 model weights to mlpack's
< KimSangYeon-DGU[> not training from the scratch
< kartikdutt18[m]> If we could do that that would be great. I'll take a look at that too.
< KimSangYeon-DGU[> Great
< KimSangYeon-DGU[> That would be nice workaround
< KimSangYeon-DGU[> You can get the weight from
< kartikdutt18[m]> About speeding up loading, I had an idea that dataloader could provide same functions as arma and when that function was called we would load and return the matrix however that would end up require a lot of templating. Eg [here](https://github.com/mlpack/mlpack/blob/30134f1ff98500cb63f50489e803392482d259d2/src/mlpack/methods/ann/ffn.hpp#L136)
< KimSangYeon-DGU[> Ok, then, can you check if it's possible to migrate the pretrained the darknet 19 model to mlapck first?
< kartikdutt18[m]> Sure
< kartikdutt18[m]> I think loading weights might be a possibility, If we get both frameworks to interact. [Loading weights in darknet](https://github.com/pjreddie/darknet/blob/f6d861736038da22c9eb0739dca84003c5a5e275/src/parser.c#L1202-L1290). I'm not sure but I think if we get the weight matrix from there and set parameters accordingly.
< sakshamb189[m]> yeah that might be a good idea to avoid the training.
< kartikdutt18[m]> I'll try this then, I'm letting it train. If it works then we can avoid training for YOLO models as well.
< kartikdutt18[m]> * I'll try this then, I'm letting the model train. If it works then we can avoid training for YOLO models as well.
< kartikdutt18[m]> * I'll try this then, I'm also letting the model train. If it works then we can avoid training for YOLO models as well.
< KimSangYeon-DGU[> The Darknet -19 model that I've run about 2.5 hours ago got 1.8 loss
< kartikdutt18[m]> Is one epoch complete?
< KimSangYeon-DGU[> Yes
< KimSangYeon-DGU[> It's on 1.5 epochs now
< kartikdutt18[m]> What was the validation accuracy?
< KimSangYeon-DGU[> I'll check that after 2 epochs done.
< KimSangYeon-DGU[> Maybe, 1hour after that will be done.
< kartikdutt18[m]> Great, Thanks.
< KimSangYeon-DGU[> :)
< saksham189Gitter> Hi @himanshupathak21061998 are you there?
< HimanshuPathakGi> Yup
< HimanshuPathakGi> How are you doing?
< saksham189Gitter> good and you?
< HimanshuPathakGi> Fine I think I will complete the implementation part of kernel svm by tonight
< saksham189Gitter> alright great and when do you think you would be able to open a PR?
< HimanshuPathakGi> @saksham189 Already opened I will tag you
< HimanshuPathakGi> For a review after completing
< HimanshuPathakGi> But right now it is not in good shape
< HimanshuPathakGi> After that I have to implement test
< saksham189Gitter> Alright I see https://github.com/mlpack/mlpack/pull/2456
< HimanshuPathakGi> You can refer to thishttps://github.com/LasseRegin/SVM-w-SMO
< HimanshuPathakGi> For implementation
< HimanshuPathakGi> But I have to make some changes to support Gaussian in this implementation
< saksham189Gitter> I think the Gaussian kernel is already implemented in `src/mlpack/core/kernels/` and you would be adding the radial basis function kernel?
< HimanshuPathakGi> Yup that's why I am adding template parameter KernelType
< HimanshuPathakGi> I will use that implementation
< saksham189Gitter> and you would be adding a radial basis function kernel right?
< HimanshuPathakGi> Radial basis kernel and Gaussian are same thing
< saksham189Gitter> oh ok.. I see
< saksham189Gitter> Have you written the blog for this week?
< HimanshuPathakGi> I want to write it after completing the implementation
< HimanshuPathakGi> So that I have something valuable to add
< saksham189Gitter> okay sure :) just make sure you share it here when you are done
< HimanshuPathakGi> Yup I will try to work tonight on it:)
< HimanshuPathakGi> I have just one concern
< saksham189Gitter> yup sure, we can discuss right now
< HimanshuPathakGi> After completing I have to compare it with libsvm
< HimanshuPathakGi> But in that they were using optimised version of smo
< HimanshuPathakGi> So, I think it will be working better than our svm
< HimanshuPathakGi> I think we should discuss this after implementation that will be better approach:)
< saksham189Gitter> I think we can first implement the basic functionality of the kernel svm and then do the comparison and discuss which optimisations if any we might want to implement. Let me know what you think.
< HimanshuPathakGi> Yup that make sense we will discuss it after implementing the basic functionality :)
< saksham189Gitter> also are you following the implementation here
< HimanshuPathakGi> Yup but I have to perform some changes in this to support Gaussian also
< HimanshuPathakGi> So the implementation will not be exact same
< saksham189Gitter> yup sure I see.
< saksham189Gitter> Let me know when the PR is ready for a review.
< saksham189Gitter> Is there anything else we should discuss?
< HimanshuPathakGi> Sure I will tag you for a review :)
< HimanshuPathakGi> Anything else you like to dicuss
< saksham189Gitter> If there are any blockers you are facing then let me know.
< HimanshuPathakGi> Yup right now I am not sure I will ask for help if I get stuck
< saksham189Gitter> alright sure. Then we will meet next time. Have a great week ahead.
< HimanshuPathakGi> Yeah until next time Have a nice week.
< HimanshuPathakGi> :)
< sakshamb189[m]> kartikdutt18: do we have a meeting right now?
< kartikdutt18[m]> Hi sakshamb189 , we do.
< sakshamb189[m]> I guess you are still training the DarkNet model, right?
< kartikdutt18[m]> Right, Let me post another update on the PR.
< sakshamb189[m]> Ok sure. Is there any improvement in the validation accuracy?
< sakshamb189[m]> I think the model might be overfitting on the CIFAR dataset since we have also reduced the training size.
< kartikdutt18[m]> Both are increasing but the increase is very slow.
< kartikdutt18[m]> *Both training and validation, however they are not good enough for classification.
< sakshamb189[m]> so what's the final validation accuracy you are getting after 3 epochs?
< kartikdutt18[m]> The third epoch isn't complete yet. It's at 94%.
< kartikdutt18[m]> For the second it was a bit over 11%
< sakshamb189[m]> and the training set size was 12.5k right?
< kartikdutt18[m]> Yes.
< sakshamb189[m]> did you check the class distribution in the train set?
< sakshamb189[m]> Is it uniform?
< kartikdutt18[m]> Each class has the number of images. (1250)
< kartikdutt18[m]> * Each class has the same number of images. (1250)
< sakshamb189[m]> and I am guessing you have a uniform distribution with the validation set right?
< kartikdutt18[m]> Hmm, I don't think so. I used the mlpack's data split and I don't think that gives uniform distribution.
< kartikdutt18[m]> It randomly picks indices from the dataset.
< sakshamb189[m]> when we used the entire train set for training we got 76% accuracy on the validation set right?
< kartikdutt18[m]> On training set.
< kartikdutt18[m]> We didn't have the validation part there then. I added it in the next trial
< sakshamb189[m]> alright and right now our train accuracy is around 11% as well .
< kartikdutt18[m]> Yes.
< sakshamb189[m]> I am a bit confused since even if the model is overfitting the train accuracy should have been higher (as compared to our previous trial with the full train set) but right now it is very low.
< kartikdutt18[m]> I can try a higher learning rate, I tried 0.1, 0.01 and 0.0001 for about 300 iterations each and only 0.001 led to decrease in loss
< kartikdutt18[m]> Hmm, I don't think this version is overfitting, I think the one with the full dataset did.
< sakshamb189[m]> Did you make any changes to the implementation after that?
< kartikdutt18[m]> I tried a few things, I tried different initialization, right now I am using random since the model didn't change it's loss with a different one. Since I was trying high learning rates I also added a batch norm layer before the last linear layer.
< kartikdutt18[m]> * I tried a few things, I tried different initialization, right now I am using random since the model didn't show decrease in loss with a different one. Since I was trying high learning rates I also added a batch norm layer before the last linear layer.
< sakshamb189[m]> why do you think the previous one was overfitting while this one is not?
< kartikdutt18[m]> The darknet 19 model converged very quickly, If that was the case it should have converged again with nearly the same params. I don't training accuracy of nearly 80% can be achieved on a single pass on the dataset.
< kartikdutt18[m]> * The darknet 19 model converged very quickly, If that was the case it should have converged again with nearly the same params. I don't think training accuracy of nearly 80% can be achieved on a single pass on the dataset.
< sakshamb189[m]> but we used almost the same architecture again with a smaller training set right ?
< kartikdutt18[m]> Right.
< kartikdutt18[m]> I even tried on the full set before training on subset I wasn;t able to reproduce the results of a few hundred iterations hence I said it was overfitting.
< sakshamb189[m]> yeah but the only changes that were made in between were adding a batch-norm layer and changing some hyper-parameters right?
< kartikdutt18[m]> Yes. I also don't understand how the accuracy was achieved.
< sakshamb189[m]> IMO we can remove the batch norm layer and try training on the small test set and see if the helps to improve the validation accuracy.
< sakshamb189[m]> Let me know what you think.
< kartikdutt18[m]> Sure, but then we can't use the current weights. I can stop the training and make the change. Also I'll try experimenting with the learning rate. I know mlpack has hyper tuner class, can we use that?
< kartikdutt18[m]> and ensmallen has grid search
< sakshamb189[m]> Alright just save the current weights.
< kartikdutt18[m]> Ohk, the weights will save after three epochs in 15-20 minutes.
< kartikdutt18[m]> *5-10 mins
< sakshamb189[m]> yes so we can wait for that to finish and then restart the training :)
< kartikdutt18[m]> Sure. Does mlpack has something similar to [this](https://towardsdatascience.com/estimating-optimal-learning-rate-for-a-deep-neural-network-ce32f2556ce0) to find the right learning rate. If we can do that for a different initialization like Xavier we would have to a lot less number of epochs since initial loss is around 2.x
< kartikdutt18[m]> * Sure. Does mlpack has something similar to [this](https://towardsdatascience.com/estimating-optimal-learning-rate-for-a-deep-neural-network-ce32f2556ce0) to find the right learning rate. If we can do that for a different initialization like Xavier we would have to do a lot less number of epochs since initial loss is around 2.x
< sakshamb189[m]> What learning rate were you using previously with the full training set?
< kartikdutt18[m]> 1e-3 same as now.
< kartikdutt18[m]> It might better to keep a log book of experiments
< sakshamb189[m]> I think we can start with that and then try other's in factors of 3 or 10
< kartikdutt18[m]> Sure, I'll maintain the logs as well.
< kartikdutt18[m]> Random Initialization or Should I spend some more timing finding one for Xavier?
< sakshamb189[m]> Currently only random initialization is implemented in mlpack right?
< kartikdutt18[m]> We have random, Xavier, He and maybe more.
< sakshamb189[m]> then I guess we could go ahead with Xavier? what do you think?
< kartikdutt18[m]> Sure, If we can find a good learning rate for that we won't have to do a lot of epochs.
< kartikdutt18[m]> Also about the transferring weights from darknet framework. I'll try to develop a proof of concept for it. If it works we can avoid training completely.
< KimSangYeon-DGU[> So sorry...
< sakshamb189[m]> Alright sure. Is there anything else we should discuss?
< sakshamb189[m]> KimSangYeon-DGU: it's fine. no worries
< kartikdutt18[m]> Not really. KimSangYeon-DGU , did you get a validation accuracy?
< KimSangYeon-DGU[> It'll be done after 2 mins
< kartikdutt18[m]> Ahh nice.
< KimSangYeon-DGU[> I had a difficulty in getting accustomed to the time difference...
< kartikdutt18[m]> No worries.
< KimSangYeon-DGU[> 8000 images left
< KimSangYeon-DGU[> So far, the accuracy is 10%
< kartikdutt18[m]> On training or on validation?
< KimSangYeon-DGU[> I did 4 epochs
< KimSangYeon-DGU[> on the test dataset in cifar/test directory
< KimSangYeon-DGU[> for the 10,000 images
< kartikdutt18[m]> Hmm, I think the mlpack implementation gives the same result in nearly same number of epochs.
< KimSangYeon-DGU[> Yes, I think the image size of CIFAR-10 is so small for the Darknet-19 architecture
< kartikdutt18[m]> Would resizing help. I think that would degrade the image quality
< KimSangYeon-DGU[> Yes, it degrades the quality and at the end, it's doesn't help...
< KimSangYeon-DGU[> The validation accuracy is 10%
< KimSangYeon-DGU[> Can we find more higher resolution image classification dataset?
< KimSangYeon-DGU[> * Can we find higher resolution image classification dataset?
< KimSangYeon-DGU[> Imagenet is too much... haha 122GB, as far as I know
< kartikdutt18[m]> I am going through [this dataset](https://ai.stanford.edu/~acoates/stl10/) right now.
< KimSangYeon-DGU[> Oh, 96x96
< KimSangYeon-DGU[> If possible, 256 x 256 is better
< kartikdutt18[m]> I think this has 500 images per class right? (That line isn't clear enough).
< KimSangYeon-DGU[> Right... it needs to be checked
< kartikdutt18[m]> Also uniform of this [dataset](http://www.vision.caltech.edu/Image_Datasets/Caltech101/) would has higher resolution images
< KimSangYeon-DGU[> Oh, nice
< KimSangYeon-DGU[> How about tiny imagenet?
< kartikdutt18[m]> It has 64x64 images and 200 classes, I guess it would be more difficult than STL, right?
< KimSangYeon-DGU[> Ahh, really? I'm downloading it, so I couldn't check the resolution...
< zoq> kartikdutt18[m]: Are you guys testing on cifar10?
< kartikdutt18[m]> Yes.
< kartikdutt18[m]> KimSangYeon-DGU, Here is the [reference](https://www.kaggle.com/c/tiny-imagenet)
< zoq> kartikdutt18[m]: to get the dataset I can just uncomment the Utils::DownloadFile section in the main right?
< kartikdutt18[m]> Yes, that should work.
< KimSangYeon-DGU[> Yeah, but the dataset has small image resolution, so I think we need to check another dataset
< zoq> kartikdutt18[m]: and start within the build folder? -> bin/object_classification ?
< kartikdutt18[m]> yes
< zoq> kartikdutt18[m]: Hm, I get - "Cannot open: No such file or directory"
< kartikdutt18[m]> Did you run sudo make ?
< zoq> kartikdutt18[m]: no
< kartikdutt18[m]> It has the same steps as mlpack, first cmake ../ && sudo make && ./bin/filename
< kartikdutt18[m]> @zoq, KimSangYeon-DGU , trained darknet 19 from the darknet framework and it got about same accuracy as the version in mlpack
< zoq> kartikdutt18[m]: Yes I did cmake and make but without sudo
< zoq> I created the data folder in the source dir, and now it downloads the dataset
< kartikdutt18[m]> I think that should work too.
< zoq> kartikdutt18[m]: But now it searches for cifar10-small
< zoq> kartikdutt18[m]: But the folder I downloaded is called cifar10.
< kartikdutt18[m]> Ahh, sorry it's the dataset that I created (a subset of cifar10 with 10k images)
< zoq> Okay, I guess I can update the code to use cifar10 than.
< kartikdutt18[m]> Could you change cifar10-small to cifar10/train then
< zoq> Is there anything else I should update before I run training?
< kartikdutt18[m]> I am pretty sure that should work.
< kartikdutt18[m]> Let me check again
< zoq> kartikdutt18[m]: Yeah, looks good.
< zoq> kartikdutt18[m]: Was just curious if the code I use the the same you use as well, to produce the graphs.
< kartikdutt18[m]> I just dump the output in a text file and read them to get the loss. What about you?
< zoq> kartikdutt18[m]: I guess I'm at the augmentation step.
< zoq> kartikdutt18[m]: "Loading Dataset!" is the last output I see.
< kartikdutt18[m]> It should take around 10 minutes to load the dataset. Augmentation is almost instant I guess.
< zoq> kartikdutt18[m]: Now "Optimizer Created, Starting Training!"
< kartikdutt18[m]> Woah, that was very fast.
< KimSangYeon-DGU[> kartikdutt18: https://github.com/fastai/imagenette and it has about 900 images per class
< kartikdutt18[m]> Did it load all images in 3 minutes?
< zoq> kartikdutt18[m]: I think so, I can time it if you like.
< kartikdutt18[m]> I think I just did. It takes 10 minutes on mine.
< kartikdutt18[m]> KimSangYeon-DGU, Nice it has benchmarks for various epochs
< zoq> kartikdutt18[m]: And I guess you report the loss after the 998 iterations?
< KimSangYeon-DGU[> Yeah, let me check with it while @zoq looks into the mlpack Darknet implementation.
< kartikdutt18[m]> I gave the output after each epoch
< kartikdutt18[m]> It's not at like 100 iteration or so right?
< KimSangYeon-DGU[> zoq: For your information, I've checked the Darknet-19 validation accuracy for CIFAR-10 in the Darknet framework, not mlpack.
< zoq> KimSangYeon-DGU[: Ahh
< KimSangYeon-DGU[> It was 10% accuracy
< KimSangYeon-DGU[> So, we thought the CIFAR-10 doesn't fit the darknet-19 architecture
< KimSangYeon-DGU[> I also checked with smaller model in the Darknet framework, the val accuracy is high about 73%
< zoq> KimSangYeon-DGU[: I see, haven't looked into the code too much, but did you scale the input to be between [0,1]?
< KimSangYeon-DGU[> No
< zoq> I guess that would be useful.
< kartikdutt18[m]> Doesn't the same script on the website handle scaling. I guess only the model was changed in that script.
< zoq> kartikdutt18[m]: What script?
< kartikdutt18[m]> I think the tutorial [here](https://pjreddie.com/darknet/train-cifar/)
< zoq> Currently looking at: https://github.com/thtrieu/darkflow/issues/38 not sure that changed.
< zoq> For mlpack the input should be in the range between [0,1], not sure there is some internal normization step in the darknet lib?
< kartikdutt18[m]> Hmm, I am also not if darknet handles it internally. In the object detection script, I used minmaxScaler on images
< KimSangYeon-DGU[> I'm not sure as well. It needs to be checked
< KimSangYeon-DGU[> zoq: As an alternative way, we thought the possibility of migration from Darknet's pretrained weights to mlpack. Does that make sense to you?
< zoq> KimSangYeon-DGU[: Yes sounds like a good idea to me, I guess we "just" have to make sure the format is the same.
< zoq> Personally I would work on the pre-trained model option first, I guess that includes images scaling as well.
< KimSangYeon-DGU[> Ok, what do you think kartikdutt18?
< kartikdutt18[m]> I will try developing POC for pretrained weights.
< KimSangYeon-DGU[> Ok, if we succeeded to do that, we can do in other models as we said :)
< KimSangYeon-DGU[> * Ok, if we succeeded to do that, we can also apply it to other models as we said :)
< walragatver[m]> jeffin143: Hi.
< jeffin143[m]> walragatver: hi
< jeffin143[m]> @kimsangyeon-dgu:matrix.org: @kartikdutt18:matrix.org are you done with your meet :)
< jeffin143[m]> If not we can wait
< KimSangYeon-DGU[> Yes, we're done :) thanks for asking
< KimSangYeon-DGU[> kartikdutt18: thanks for the meeting and please ping me if you have anything you want to discuss!
< KimSangYeon-DGU[> * jeffin143: Yes, we're done :) thanks for asking
< jeffin143[m]> KimSangYeon-DGU (@kimsangyeon-dgu:matrix.org): :)
< jeffin143[m]> walragatver: anything on your list
< jeffin143[m]> I don't have anything
< jeffin143[m]> I have updated the testing pr with some test , have a look
< kartikdutt18[m]> Sure. walragatver Please continue, thanks
< jeffin143[m]> Also if possible need your review on image logging pr too
< jeffin143[m]> Also I have updated the blog for this week :)
< jeffin143[m]> That's everything from my side
< walragatver[m]> jeffin143: Sorry I got deviated
< walragatver[m]> I will give you review soon.
< jeffin143[m]> walragatver: no issues :) all good
< jeffin143[m]> I have opened two pr for the reason so that work doesn't stop
< walragatver[m]> jeffin143: What's the plan ahead? What would you be implementing next?
< jeffin143[m]> walragatver: if we are done with image and testing
< jeffin143[m]> I will go for text and audio next
< jeffin143[m]> I am planning to complete image and testing pr before first phase ends
< walragatver[m]> jeffin143: And I saw your blog. I think from this year participants are allowed to write blogs anywhere
< walragatver[m]> <jeffin143[m] "I will go for text and audio nex"> Okay
< jeffin143[m]> > jeffin143: And I saw your blog. I think from this year participants are allowed to write blogs anywhere
< jeffin143[m]> Yes, but I guess that was true every year, we just went with irc last year :)
< jeffin143[m]> Sry the blog repo*
< RyanBirminghamGi> jeffin143: I'll also give another pass at reviews soon!
< jeffin143[m]> <jeffin143[m] "Sry the blog repo*"> Ryan Birmingham (Gitter): sure :)
< walragatver[m]> jefffin143: Okay, you are sure that this time we are allowed to use the blog repo?
< jeffin143[m]> I guess so
< walragatver[m]> Because no else is using it. And I think there was no mention of the blog repo in the introductory mails this time.
< walragatver[m]> It might happen that we stop maintaining it. So just confirm it.
< jeffin143[m]> May be I can send it through mailing list too
< jeffin143[m]> walragatver: may be I can send it through mailing list too
< walragatver[m]> jeffin143: It's fine I would say to continue with blog. Just change the location if something happens. Don't use mailing list because it's a communication medium.
< jeffin143[m]> I will confirm it with Ryan and let you know
< walragatver[m]> jeffiin143: Are you going to also add the graph support in the library?
< jeffin143[m]> walragatver: we don't follow the graph convention in mlpack , right ?
< walragatver[m]> Graph in the sense histogram etc.
< jeffin143[m]> Yes histogram and pr curve
< jeffin143[m]> Text audio
< jeffin143[m]> Embedding
< jeffin143[m]> 5 more things are there in my list
< walragatver[m]> Okay
< walragatver[m]> The path ahead would be quite smooth now as the CI and testing implementation is over.
< jeffin143[m]> walragatver: yes , I am sure we can speed a little bit :)
< jeffin143[m]> I can write more tutorials then may be if I am left with some time
< walragatver[m]> jeffin143: Have you given any thoughts on callbacks? Any implementatiion idea or something?
< jeffin143[m]> walragatver: no , I will try to sketch it this week
< jeffin143[m]> May be then we can schedule a meet with zoq to get input as well ??
< walragatver[m]> jeffin143: And what about your joining. When would you be joining the firm?
< jeffin143[m]> If a user Wants to log something else he has to create it as a custom function and pass it the callback
< walragatver[m]> jeffin143: I am not sure about the tf callbacks i will check it out.
< walragatver[m]> jeffin143: If it's all about accuracy and loss then it would be quite easy to implement.
< jeffin143[m]> > jeffin143: And what about your joining. When would you be joining the firm?
< jeffin143[m]> The contacted me and told me that hiring are paused they wouldn't be offering me a full time job
< jeffin143[m]> They said they will try a contract offer not to leave me in mid
< jeffin143[m]> So I am unsure
< jeffin143[m]> > jeffin143: Really? That's slightly troublesome.
< jeffin143[m]> Umm , tough situations , can't complain
< jeffin143[m]> @walragatver:matrix.org: did you join any org ?
< walragatver[m]> <jeffin143[m] "@walragatver:matrix.org: did you"> jeffin143: Nope I might join soon
< jeffin143[m]> walragatver: Probably it would be a work from home onbaording
< jeffin143[m]> :) good luck with it
< walragatver[m]> <jeffin143[m] "walragatver: Probably it would b"> jeffin143: Yeah most probably
< walragatver[m]> <jeffin143[m] ":) good luck with it"> Thanks. Lets see
< jeffin143[m]> walragatver: Thanks for the time ,
< jeffin143[m]> I will take a leave :) good night
< walragatver[m]> Yeah take care.
< RyanBirminghamGi> jeffin143: walragatver: Sorry I didn't say much of anything today; have a good night!
< walragatver[m]> I hope birm doesn't have anything to discuss
< walragatver[m]> <RyanBirminghamGi "jeffin143: walragatver: Sorry I "> No problem birm take care
< jeffin143[m]> Ryan Birmingham (Gitter): no issues :)
< jeffin143[m]> Take care , have a good week ahead
< zoq> kartikdutt18[m]: Btw. running with BatchSize=1 and looks like the loss is going down with each step currently at 0.00925315.
< kartikdutt18[m]> What was the initial loss?
< kartikdutt18[m]> Is an epoch almost done ?
< kartikdutt18[m]> * Also, is an epoch almost done ?
< zoq> kartikdutt18[m]: Can't remember I'm at: 1000/7977
< kartikdutt18[m]> Ahh, That's nice.
< zoq> kartikdutt18[m]: Let me print each step.
< kartikdutt18[m]> Sure, Thanks. I am stopping my model now as discussed.
< kartikdutt18[m]> My loss at start of the 4 th epoch is nearly 4.5
ImQ009 has quit [Quit: Leaving]
< KimSangYeon-DGU[> zoq: Thanks!
< HimanshuPathakGi> Hey @zoq @rcurtin How can we find dot product of two arma::mat??
< zoq> HimanshuPathakGi: A * A.t() maybe I miss something?
< HimanshuPathakGi> > `zoq on Freenode` Himanshu Pathak (Gitter): A * A.t() maybe I miss something?
< HimanshuPathakGi> Yup got it :)
< nishantkr18[m]> Hey everyone! Here's a 2 min read on my last two week's progress. https://nishantkr18.github.io/GSoC-2020-mlpack/week-04-and-05/week-04-and-05.html Kindly have a look :)