You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@spark.apache.org by nkd <ka...@gmail.com> on 2014/04/06 22:05:35 UTC

Re: any work around to support nesting of RDDs other than join

It worked when I converted the nested RDD to an array

------------------------------
case class TradingTier(tierId:String, lowerLimit:Int,upperLimit:Int ,
transactionFees:Double) 
    
   //userTransactions Seq[(accountId,numTransactions)] 
   val userTransactionsRDD =
sc.parallelize(Seq((id1,2),(id2,4),(id3,1),(id4,3))) 

   val transactionTiersRDD =  
sc.parallelize(Seq(TradingTier("tier1",0,2,9.00),TradingTier("tier2",1,4,3.00),TradingTier("tier3",3,5,5.00))) 

  * val transactionTiersArray =   transactionTiersRDD.toArray*

val userTransactionFees = userTransactionsRDD.flatMap( x => 
        transactionTiersArray.filter(y=>(y.upperLimit > x._2 && y.lowerLimit
< x._2)).map(_.transactionFees)) 
   val aggregateTransactionTiers2 =
userTransactionsRDD.zip(userTransactionFees) 
------------------------------




--
View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/any-work-around-to-support-nesting-of-RDDs-other-than-join-tp3816p3820.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.