Tag: Spark_Interview

PySpark @ Freshers.in

PySpark : Inserting row in Apache Spark Dataframe.

In PySpark, you can insert a row into a DataFrame by first converting the DataFrame to a RDD (Resilient Distributed…

Continue Reading PySpark : Inserting row in Apache Spark Dataframe.
PySpark @ Freshers.in

PySpark : How to write Scala code in spark shell ?

To write Scala code in the Spark shell, you can simply start the Spark shell by running the command “spark-shell”…

Continue Reading PySpark : How to write Scala code in spark shell ?
PySpark @ Freshers.in

PySpark : What happens once you do a spark submit command ?

When you submit a Spark application using the spark-submit command, a series of steps occur to start and execute the…

Continue Reading PySpark : What happens once you do a spark submit command ?
PySpark @ Freshers.in

Installing Apache Spark standalone on Linux

Installing Spark on a Linux machine can be done in a few steps. The following is a detailed guide on…

Continue Reading Installing Apache Spark standalone on Linux
PySpark @ Freshers.in

What is GC (Garbage Collection) time in Spark UI ?

In the Spark UI, GC (Garbage Collection) time refers to the amount of time spent by the JVM (Java Virtual…

Continue Reading What is GC (Garbage Collection) time in Spark UI ?
PySpark @ Freshers.in

PySpark : How do I read a parquet file in Spark

To read a Parquet file in Spark, you can use the spark.read.parquet() method, which returns a DataFrame. Here is an…

Continue Reading PySpark : How do I read a parquet file in Spark