ChanServ changed the topic of #mlpack to: "mlpack: a fast, flexible machine learning library :: We don't always respond instantly, but we will respond; please be patient :: Logs at http://www.mlpack.org/irc/
< rcurtin[m]>
unfortunately I didn't see it either; by the time I made it outside, Jupiter and Saturn had already set :(
< AyushSingh[m]>
zoq are you referring to adding seq2seq model to the models repository?
< GauravGhati[m]>
Hey ryan I opened a pr few
< GauravGhati[m]>
Hey #2769 few days back. please have a look whenever you have time.
ib07 has quit [Quit: No Ping reply in 180 seconds.]
ib07 has joined #mlpack
kristjansson has joined #mlpack
ib07 has quit [Ping timeout: 256 seconds]
ib07 has joined #mlpack
Samyak has joined #mlpack
< zoq>
AyushSingh[m]: Yes, that's what I had in mind.
< Samyak>
Hi, I am working on parallelization of algorithms implemented in mlpack using MPI(https://en.wikipedia.org/wiki/Message_Passing_Interface) It would give improvement in runtime performance. Has anyone else worked on parallelization before?
tarunjarvis5Gitt has joined #mlpack
< tarunjarvis5Gitt>
I am new to open source I was suggested to go through the goodfirstissue but as I am new I can hardly understand it can someone walk me through
< rcurtin[m]>
Hi Samyak, I would recommend against using MPI directly. It will make the implementations too complex. Plus, mlpack is not really designed for clusters or situations where there are multiple distinct systems. I would suggest focusing on OpenMP because even after OpenMP support is added, the code is still understandable to people who do not know OpenMP well (that is definitely not the case with MPI)
< Samyak>
MPI gave huge performance improvement on my machine. MPI would use all cores of a machine whereas OpenMP uses multithreading in a shared memory setting(single core). I would provide proper documentation wherever MPI is used.
< rcurtin[m]>
A shared memory setting is the setting that the vast majority of people will run mlpack in: single node, multicore. mlpack is not really built for the multinode setting, and even if you did adapt some of the algorithms to use MPI, often the algorithmic strategy for machine learning algorithms in a multinode context needs to be vastly different due to communication overhead. If you are looking to implement MPI-based
< rcurtin[m]>
machine learning algorithms, perhaps finding an explicitly distributed machine learning library might be a better choice?
ib07 has quit [Ping timeout: 272 seconds]
tarunjarvis5 has joined #mlpack
< tarunjarvis5>
I am new to open source I was suggested to go through the goodfirstissue but as I am new I am having difficulty understanding it can someone walk me through
Samyak has quit [Remote host closed the connection]
tarunjarvis5 has quit [Remote host closed the connection]
< zoq[m]>
Connection lost...
< rcurtin[m]>
oops :(
< _slack_mlpack_U0>
What happened 😂
< AyushSingh[m]>
tarunjarvis5 - You can select the issue of your choice, read more about it(it's conversation thread and also through external sources), look in the thread if any PR related to it is merged, get a basic understanding of what is being done over there, look how people have contributed to that PR in the past and what all is left to be done and how can you contribute in it.
< AyushSingh[m]>
zoq , okay, I will look into how to implement it.
< zoq>
rcurtin[m]: About the NF monthly update, maybe it makes sense to put that on git so we can add things, and just take what's there once Walter asked.
< rcurtin[m]>
zoq: sure, that seems like a great idea; where should we put it?
< rcurtin[m]>
we could just take HISTORY.md perhaps, but that won't catch any PRs that aren't yet merged or things that don't have to do with the codebase
< zoq>
HISTORY.md also only covers mlpack/mlpack
< zoq>
I guess a wiki page or some new repo works?
< rcurtin[m]>
yeah, I suppose either would be just fine
himanshu_pathak[ has quit [Quit: Idle for 30+ days]
himanshu_pathak[ has joined #mlpack
tarunjarvis5 has joined #mlpack
tarunjarvis5 has quit [Remote host closed the connection]
The_LoudSpeaker has quit [Quit: Leaving bye!]
The_LoudSpeaker has joined #mlpack
ib07 has joined #mlpack
ekdnam has joined #mlpack
ekdnam has quit [Remote host closed the connection]
< zoq[m]>
Let us know if there is anything we can clarify.
qur70 has quit [K-Lined]
ib07 has quit [Ping timeout: 240 seconds]
ib07 has joined #mlpack
ib07 has quit [Ping timeout: 240 seconds]
ib07 has joined #mlpack
ib07 has quit [Ping timeout: 240 seconds]
ib07 has joined #mlpack
ib07 has quit [Ping timeout: 256 seconds]
ib07 has joined #mlpack
< rcurtin[m]>
@dkipke interesting. DBSCAN will sometimes not assign points to clusters---and as a result the returned assignment will be SIZE_MAX (or its equivalent in Go); do you think that is what is happening with your data?
ib07 has quit [Ping timeout: 240 seconds]
< rcurtin[m]>
hmm... what happens if you try to scale k-means or mean shift in the same way?