Jubatus Explained

Jubatus
Collapsible:yes
Developer:Nippon Telegraph and Telephone & Preferred Infrastructure
Latest Release Version:0.4.3
Operating System:Linux
Programming Language:C++
Genre:machine learning
License:GNU Lesser General Public License 2.1

Jubatus is an open-source online machine learning and distributed computing framework developed at Nippon Telegraph and Telephone and Preferred Infrastructure. Its features include classification, recommendation, regression, anomaly detection and graph mining.It supports many client languages, including C++, Java, Ruby and Python.It uses Iterative Parameter Mixture[1] [2] for distributed machine learning.

Notable Features

Jubatus supports:

Notes and References

  1. Ryan McDonald, K. Hall and G. Mann, Distributed Training Strategies for the Structured Perceptron, North American Association for Computational Linguistics (NAACL), 2010.
  2. Gideon Mann, R. McDonald, M. Mohri, N. Silberman, and D. Walker, Efficient Large-Scale Distributed Training of Conditional Maximum Entropy Models, Neural Information Processing Systems (NIPS), 2009.
  3. Koby . Crammer. Ofer. Dekel. Ofer Dekel (researcher) . Shai. Shalev-Shwartz. Yoram . Singer . Online Passive-Aggressive Algorithms . Proceedings of the Sixteenth Annual Conference on Neural Information Processing Systems (NIPS) . 2003.
  4. Koby Crammer and Yoram Singer. Ultraconservative online algorithms for multiclass problems. Journal of Machine Learning Research, 2003.
  5. Koby Crammer, Ofer Dekel, Joseph Keshet, Shai Shalev-Shwartz, Yoram Singer, Online Passive-Aggressive Algorithms. Journal of Machine Learning Research, 2006.
  6. Mark Dredze, Koby Crammer and Fernando Pereira, Confidence-Weighted Linear Classification, Proceedings of the 25th International Conference on Machine Learning (ICML), 2008
  7. Koby Crammer, Mark Dredze and Fernando Pereira, Exact Convex Confidence-Weighted Learning, Proceedings of the Twenty Second Annual Conference on Neural Information Processing Systems (NIPS), 2008
  8. Koby Crammer, Mark Dredze and Alex Kulesza, Multi-Class Confidence Weighted Algorithms, Empirical Methods in Natural Language Processing (EMNLP), 2009
  9. Koby Crammer, Alex Kulesza and Mark Dredze, Adaptive Regularization Of Weight Vectors, Advances in Neural Information Processing Systems, 2009
  10. Koby Crammer and Daniel D. Lee, Learning via Gaussian Herding, Neural Information Processing Systems (NIPS), 2010.