In this chapter, we looked at executing jobs in Sahara by examining the running job workflow on top of OpenStack. The EDP project in OpenStack with the last releases becomes a great flavor to run complex jobs on a Hadoop cluster in no time with the support of many other plugins that have been cited in this chapter.
This chapter also covered how to launch a Spark cluster and execute a simple job using the Sahara REST API. By the end of the chapter, an example of a real-world scenario was discussed and it was explained how Sahara can simplify the scalability challenge. Running more complex jobs within tons of unstructured data will undoubtedly need a simple and easy approach to tackle future data growth issues. The chapter has demonstrated how Sahara is capable of accomplishing such a necessity.
The next chapter will go in depth on more advanced options that Sahara offers for a more sophisticated data processing environment.