Reading and Writing Parquet nested datatype file using Pyspark

write a pyspark job which should read a parquet file which has nested datatypes & values( records) and change the one of the column value with xxx and write into a new parquet file.

so the actual source file and newly created file should be same with the small change ( changed to xxx for of the row value for one column), the new parquet file should same as the original file (schema, no of records, order of records) with the changed one of the column value 'xxx'

Note:- logic should be dynamic ,parquet file schema will not be the same all the time.....our code should read the parquet file schema dynamically and and create the parquet file with changed data ( xxx) ....the rows, schema and columns should be same

Evner: PySpark

Se mere: playing wmv file using directx, saving grid data xml file using extjs, mysql xml file using php, reading writing xml java, read xml file using net insert sql server, christmas ideas using reading writing, reading csv file using adodb, reading csv file using ado, simple code reading xml file using vbnet 2005, peoplesoft writing file using file layout, reading data excel file using, opening writing excel file using php, vbnet reading csv file using adonet, reading file using ftp vba, reading text file using adonet, writing excel file using, reading xls file using php, reading and writing files in sql server using t-sql, implement a program in reading and writing from a file using multiple threads in c, implement a program in reading and writing from a file using multiple threads.

Om arbejdsgiveren:
( 3 bedømmelser ) Mountain House, United States

Projekt ID: #31013380

3 freelancere byder i gennemsnit $7/timen for dette job


I read your project description carefully. I am bidding on your project because I am very much familiar with Python, Pyspark and Parquet. I am an experienced Data Scientist and Machine Learning Engineer. Data Visualiz Flere

$5 USD / time
(0 bedømmelser)

I will work on this , if I am given opportunity. So basically I would like to discuss few things before I am alloted to this

$8 USD / time
(0 bedømmelser)

I do have 6 years of IT experience with Python, pyspark, Spark, SQL, ETL, data analysis and data engineering. I do have exposure to Azure and Aws cloud platforms. Though, i am quite new to this platform but i can assu Flere

$7 USD / time
(0 bedømmelser)