I am relatively new to Apache Spark and Python and was wondering how to get the size of a RDD. I have a RDD that looks like this:
[[‘ID: 6993.1066',
'Time: 15:53:43',
'Lab: West',
'Lab-Tech: Nancy McNabb, ',
'\tBob Jones, Harry Lim, ',
'\tSue Smith, Will Smith, ',
'\tTerry Smith, Nandini Chandra, ',
]]
Is there a method or function in pyspark that can give the size how many tuples in a RDD? The one above has 7.
Scala has something like: myRDD.length.