Spark is a unique framework for big data analytics which gives one unique integrated API by developers for the purpose of data scientists and analysts to perform separate tasks. It supports a wide range of popular languages like Python, R, SQL, Java and Scala. Apache Spark main aim is to provide hands-on experience to create real-time Data Stream Analysis and large-scale learning solutions for data scientists, data analysts and software developers. Spark is a cluster computer framework. It is an open source and lighting fast computer designer. It is used or extends the MapReduce model. It was introduced by Apache framework. It enables a computing solution which is scalable. Also, it is flexible and cost-effective. It is used for the speed up the Hadoop computational computing software process. Spark has its own cluster management, hence it uses the Hadoop for the storage purpose.