What would be the best way to use Spark and neutral networks (especially RNN LSTM) ?
I think it would be possible by "tool"-combination:
Pyspark + anaconda + pandas + numpy + keras + tensorflow + scikit
But what about scalability and usability by Spark (pyspark) ?
How compatible are data structures (for example dataframe) between Spark and other "tools" ?
And it is possible convert them between different "tools" ?