I have a question similar but somehow opposite to what was considered here:
We are also trying to use SCIFIO API with Apache Spark. We would like to perform some operations in parallel on large numbers of independent images. I managed to use solutions provided in the topic above to read an image from an array of bytes, so I’m not dependent on the file system when reading - in our case files may come from HDFS (Hadoop File System), https, or even Microsoft Azure specific storage.
Previously we developed procedures to store arrays of bytes representing image files into these systems. However, the only way I see now to save an image with SCIFIO procedures is to save it to a local file. I could create a temporary file with them, read it into an array of bytes and then save it to these special systems but this is too much of a workaround. Am I missing something? Is it possible to save an image into an array of bytes (or a stream) instead of into a file?