ChanServ changed the topic of #mlpack to: "mlpack: a fast, flexible machine learning library :: We don't always respond instantly, but we will respond; please be patient :: Logs at http://www.mlpack.org/irc/
gtank___ has quit [Read error: Connection reset by peer]
gtank___ has joined #mlpack
< metahost> zoq: could you please review ensmallen#149 once. I have added a comment.
< jenkins-mlpack2> Project docker mlpack nightly build build #586: UNSTABLE in 2 hr 53 min: http://ci.mlpack.org/job/docker%20mlpack%20nightly%20build/586/
ImQ009 has joined #mlpack
SriramSKGitter[m has joined #mlpack
< SriramSKGitter[m> When I run bin/mlpack I get the following:
< SriramSKGitter[m> bin/mlpack_test: symbol lookup error: bin/mlpack_test: undefined symbol: _ZN6mlpack4tree10CosineTreeC1ERKN4arma3MatIdEE
< SriramSKGitter[m> How do I go about debugging this?
< SriramSKGitter[m> *bin/mlpack_test not bin/mlpack
xiaohong has joined #mlpack
xiaohong has quit [Remote host closed the connection]
xiaohong has joined #mlpack
xiaohong has quit [Remote host closed the connection]
xiaohong has joined #mlpack
xiaohong has quit [Ping timeout: 260 seconds]
xiaohong has joined #mlpack
xiaohong has quit [Remote host closed the connection]
launchpad5682 has joined #mlpack
xiaohong has joined #mlpack
dantraz has joined #mlpack
xiaohong has quit [Remote host closed the connection]
dantraz has quit [Remote host closed the connection]
xiaohong has joined #mlpack
xiaohong has quit [Ping timeout: 260 seconds]
xiaohong has joined #mlpack
xiaohong has quit [Ping timeout: 260 seconds]
xiaohong has joined #mlpack
xiaohong has quit [Ping timeout: 260 seconds]
launchpad5682 has quit [Ping timeout: 260 seconds]
xiaohong has joined #mlpack
xiaohong has quit [Ping timeout: 260 seconds]
Soham has joined #mlpack
xiaohong has joined #mlpack
Soham has quit [Remote host closed the connection]
xiaohong has quit [Ping timeout: 260 seconds]
xiaohong has joined #mlpack
xiaohong has quit [Ping timeout: 260 seconds]
xiaohong has joined #mlpack
xiaohong has quit [Ping timeout: 260 seconds]
xiaohong has joined #mlpack
xiaohong has quit [Ping timeout: 260 seconds]
PrinceGuptaGitte has joined #mlpack
< PrinceGuptaGitte> Hi, I'm new to open source but love machine learning. I wanted to start contributing to MLpack, I built it from source in ubuntu, everything is working fine except this line `data::Save("cov.csv", cov, true);`
< PrinceGuptaGitte> I get this error, I've been trying to figure this out but couldn't.
xiaohong has joined #mlpack
xiaohong has quit [Ping timeout: 260 seconds]
xiaohong has joined #mlpack
xiaohong has quit [Ping timeout: 260 seconds]
xiaohong has joined #mlpack
< kartikdutt18Gitt> Hi @rcurtin , @zoq I have implemented mish activation function #2156, I already had a PR open what should I do, Currently changes made are visible in the previous [PR:](https://github.com/mlpack/mlpack/pull/2126).Thanks.
< rcurtin> I don't understamd what the question is
xiaohong has quit [Ping timeout: 260 seconds]
yash71 has joined #mlpack
yash71 has quit [Remote host closed the connection]
< kartikdutt18Gitt> Hi,I already have a PR open regarding another issue and I am unable to open a new PR so should stash changes of previous PR, close it and send a new PR or should I create a new branch with changes of mish function only?
< kartikdutt18Gitt> The changes can be viewed in the PR #2126 .
< rcurtin> kartikdutt18Gitt: it sounds to me like you should use a different branch on your mlpack fork for the second set of changes, and then you can open two PRs
xiaohong has joined #mlpack
< kartikdutt18Gitt> Ohh thanks, I will create a new branch and send another pull request. Thanks.
xiaohong has quit [Ping timeout: 260 seconds]
< rcurtin> sounds good :)
xiaohong has joined #mlpack
xiaohong has quit [Ping timeout: 260 seconds]
< kartikdutt18Gitt> Hi @rcurtin, I have created a new pull request with only the second set of changes.
SachaD has joined #mlpack
rahul has joined #mlpack
< rahul> hi everyone, I m Rahul and I want to contribute in this organisation for the upcoming gsoc 2020,I m good at c++,java & do competitive programming so can anyone help me out how can I start with solving bugs & contribute as well
Param-29Gitter[m has joined #mlpack
< Param-29Gitter[m> Hello, I would like to know what is meaning of relevant tickets mentioned on https://github.com/mlpack/mlpack/wiki/SummerOfCodeIdeas
xiaohong has joined #mlpack
< SachaD> Zoq, my minimum example:
< SachaD> #include <mlpack/core.hpp>
< SachaD> #include <mlpack/methods/ann/ffn.hpp>
< SachaD> #include <mlpack/methods/ann/layer/layer.hpp>
< SachaD> #include <mlpack/methods/ann/loss_functions/mean_squared_error.hpp>
< SachaD> #include <armadillo>
< SachaD> #include <omp.h>
< SachaD> using namespace std;
< SachaD> int main(int argc, char* argv[])
< SachaD> {
< SachaD> arma::mat dataset, trainData, outLabels;
< SachaD> mlpack::ann::FFN<> model;
< SachaD> static const int maxWordLength = 31;
< SachaD> mlpack::data::Load("en-dic-utf16-codes.csv", dataset, true);
< SachaD> if (dataset.n_rows != maxWordLength){
< SachaD> return -1;
< SachaD> }
< SachaD> trainData = dataset.submat(0, 0, dataset.n_rows - 1, dataset.n_cols - 1);
< SachaD> outLabels = dataset.submat(0, 0, dataset.n_rows - 1, dataset.n_cols - 1);
< SachaD> outLabels.for_each( [](arma::mat::elem_type& val) { val += 1; } );
< SachaD> model.Add< mlpack::ann::Linear<> >(maxWordLength, maxWordLength);
< SachaD> model.Add< mlpack::ann::SigmoidLayer<> >();
< SachaD> model.Add< mlpack::ann::Linear<> >(maxWordLength, maxWordLength);
< SachaD> model.Add< mlpack::ann::LogSoftMax<> >();
< SachaD> model.Train(trainData, outLabels);
< SachaD> return 0;
< SachaD> }
< SachaD> link to file "en-dic-utf16-codes.csv": https://yadi.sk/d/RxhD2xALODmy2w
< SachaD> P.S. mlpack works with double data type. Can it work with float or even unsigned short? - they are faster end less expensive.
xiaohong has quit [Ping timeout: 260 seconds]
rahul has quit [Ping timeout: 260 seconds]
garg has joined #mlpack
garg has quit [Remote host closed the connection]
xiaohong has joined #mlpack
xiaohong has quit [Ping timeout: 260 seconds]
xiaohong has joined #mlpack
xiaohong has quit [Ping timeout: 260 seconds]
< kartikdutt18Gitt> Hi SachaD, double has 52 mantissa bits where as float has 23 bits so it double offers more precision in general, Also I think by default PyTorch also requires tensors to be of type float (generally).
nishantkr18 has joined #mlpack
xiaohong has joined #mlpack
xiaohong has quit [Ping timeout: 260 seconds]
nishantkr18 has quit [Remote host closed the connection]
nishantkr18 has joined #mlpack
< PrinceGuptaGitte> Which library is is responsible when we save a armadillo matrix to a CSV?
< PrinceGuptaGitte> Like this: `data::Save("cov.csv", cov, true);`
nishantkr18 has quit [Remote host closed the connection]
< SachaD> kartikdutt18Gitt, yes, but I see the lections on youtube about deep learing and there was told that gpu makes faster multiplications of shorter datatypes. Thats why TensorFlow uses tf.float16 and tf.uint16 for faster speed and less memory. Also GPUs optimized to multiple 16 and 32 bit matices (double is 64 bit). 16 bit is twice more speed/less memory than 32 bit values. I am a c++ programmer, thats why I see in mlpack side, but if
< SachaD> Tesorflow can operate smaller data types it can be faster and more effective than mlpack or I missed something (Yes, c++ is faster than Python, but if there a large amount of data and calculations it can be beated with more optimized calculations and data storage)?
xiaohong has joined #mlpack
xiaohong has quit [Ping timeout: 245 seconds]
xiaohong has joined #mlpack
< zoq> SachaD: Not every method in mlpack is templated, so that you can use arma::Mat<float> instead of arma::Mat<double>, I guess a hacky solution is to use: typedef double float;.
ImQ009 has quit [Quit: Leaving]
xiaohong has quit [Ping timeout: 260 seconds]
< zoq> SachaD: 8bit integer is the same, but without a good model design this isn't going to work, you will see a speedup but a drop in accuracy; tf provides some support to quantize the model, but that doesn't work for each model.
< zoq> Param-29: Hello, sometimes we have issues that might be a good starting point to get familair with a certain project.
< zoq> rahul: Hello there, https://www.mlpack.org/gsoc.html should answer your questions, let us know if you have any further questions.
< sreenik[m]> SachaD: The tensor cores in modern Nvidia GPUs (Turing and Volta probably) have some extremely capable Tensor cores in addition to Cuda cores which only work with fp16 and fp8 (not with fp32). Is that what you are emphasizing on?
< zoq> PrinceGupta: In this case that would be armadillo.
xiaohong has joined #mlpack
xiaohong has quit [Ping timeout: 245 seconds]
xiaohong has joined #mlpack
xiaohong has quit [Ping timeout: 260 seconds]
< SachaD> sreenik[m], yes.
SachaD has quit [Quit: Konversation terminated!]
xiaohong has joined #mlpack
xiaohong has quit [Ping timeout: 260 seconds]
xiaohong has joined #mlpack
xiaohong has quit [Ping timeout: 260 seconds]
< sreenik[m]> SachaD: Yes, popular frameworks use a mixed precision to speed up stuff and somewhat preserve accuracy but in the end the low level BLAS routines used are independent of whether we are using python or c++. Currently with mlpack/armadillo, we use Nvblas for gpu based operations (if you enable it) but the data transfer overheads deter most of the speed gains. Since Nvblas cannot handle all sorts of operations that Cublas
< sreenik[m]> can, it directs some of them to the cpu (default blas routine like openblas, etc.) and some to the gpu instead of copying the entire data to the gpu memory at once. I suppose Tensorflow, etc. directly make use of Cublas and direct all operations to the gpu which makes the operation very fast. As far as mlpack is concerned, Bandicoot is in development (built on top of armadillo) and would address these issues. zoq correct
< sreenik[m]> me if I am wrong.
xiaohong has joined #mlpack
xiaohong has quit [Ping timeout: 260 seconds]
< rcurtin> SachaD: sreenik[m]: I'm agreed on everything that's written here. realistically, TensorFlow is C++ under the hood. at the moment, mlpack isn't able to make use of GPUs except for early Bandicoot support
< rcurtin> essentially we have been waiting on a "nice" GPU linear algebra library instead of hand-maintaining awful CUDA code, etc.
< rcurtin> there is still some extra overhead that running TF is likely to incur, but I of course couldn't predict exactly how mlpack+bandicoot would fare against TF (or PyTorch). we will find out soon enough once bandicoot is ready, I think
xiaohong has joined #mlpack
xiaohong has quit [Ping timeout: 260 seconds]
lozhnikov has quit [Ping timeout: 265 seconds]
lozhnikov has joined #mlpack