# importing module
import pyspark
# importing sparksession from pyspark.sql module
from pyspark.sql import SparkSession
# creating sparksession and giving an app name
spark = SparkSession.builder.appName('sparkdf').getOrCreate()
# list of employee data with 10 row values
data =[["1","sravan","IT",45000],
["2","ojaswi","IT",30000],
["3","bobby","business",45000],
["4","rohith","IT",45000],
["5","gnanesh","business",120000],
["6","siva nagulu","sales",23000],
["7","bhanu","sales",34000],
["8","sireesha","business",456798],
["9","ravi","IT",230000],
["10","devi","business",100000],
]
# specify column names
columns=['ID','NAME','sector','salary']
# creating a dataframe from the lists of data
dataframe = spark.createDataFrame(data,columns)
# display dataframe
dataframe.show()