How to convert spark RDD to mahout DRM? -


i fetching data alluxio in mahout using sc.textfile(), spark rdd. program further uses spark rdd mahout drm, therefore needed convert rdd drm. current code remains stable.

an apache mahout drm can created apache spark rdd in following steps:

  1. convert each row of rdd mahout vector
  2. zip rdd index (and swap tuple of form (long, vector)
  3. wrap rdd drm.

consider following example code:

val rdda = sc.parallelize(array((1.0, 2.0, 3.0),             ( 2.0, 3.0, 4.0),             ( 4.0, 5.0, 6.0)))  val drmrdda: drmrdd[long] = rdda.map(a => new densevector(a))                  .zipwithindex()                  .map(t => (t._2, t._1))  val drma = drmwrap(rdd= drmrdda) 

source /more info/ shameless self promotion (toward bottom): my blog


Comments

Popular posts from this blog

Command prompt result in label. Python 2.7 -

javascript - How do I use URL parameters to change link href on page? -

amazon web services - AWS Route53 Trying To Get Site To Resolve To www -