window - How many RDD in the resulting DStream of reduceByKeyAndWindow -


i working on small spark job compute stock correlation matrix dstream.

from dstream[(time, quote)], need aggregate quotes (double) time (long) among multiple rdds, before computing correlations (considering quotes of rdds)

dstream.reducebykeyandwindow{./*aggregate quotes in vectors*/..}         .foreachrdd {rdd => statistics.corr(rdd[vector])} 

to mind, solution if resulting dstream (from reducebykeyandwindow) contains 1 rdd aggregated quotes.

but not sure. how data distributed after reducebykeyandwindow? there way merge rdds in dstream?


Comments

Popular posts from this blog

gcc - MinGW's ld cannot perform PE operations on non PE output file -

timeout - Handshake_timeout on RabbitMQ using python and pika from remote vm -

c# - Search and Add Comment with OpenXML for Word -