Hello,
I need to generate a large chunk of data from multiple saved searches and loading of records and then store it in a single pdf/excel file.
The issue is if I use a Map/Reduce then it will yield multiple times and how can I continue to append to the same file created earlier during the start of the job.
How can I continue writing data from the each loaded record to this same file.
Map/Reduce will surely run multiple times over and afresh, how can I share the file ID parameter to the same script run so that 1 map/reduce produces 1 single file.
Basically I need to control in a way that each yield I know the data reached and remaining to be processes and write to the same file created during first run.
Any thoughts or suggestions are most welcome.
Note - I have done this with scheduled script in the past successfully but just want to know if map/reduce can do it better or if thats the only way.