NickName:Yunus Emrah Uluçay Ask DateTime:2022-03-13T13:56:12 How can I run PySpark on a single node and multiple node Hadoop Environment? I need a one single node and multiple node hadoop environment on docker and i need to make some analysis using PySpark on these hadoop environments. Now i am trying one single node. I pull an ubuntu image, containerized it and installed hadoop environment on this container but i confused whether spark runs on a installed hadoop environment or it needs to install its own environment which has hadoop(Maybe the sentence is complicated, is spark establish on hadoop or is spark install hadoop while its own installation?). Copyright Notice:Content Author:「Yunus Emrah Uluçay」,Reproduced under the CC 4.0 BY-SA copyright license with a link to the original source and this disclaimer.Link to original article:https://stackoverflow.com/questions/71454224/how-can-i-run-pyspark-on-a-single-node-and-multiple-node-hadoop-environment Answers