Skip to main content
 print this page

Writing Data to Amorphic Datalake

Following are some examples on how to read using amorphicutils.

1. Python shell

1.1. Writing CSV data

1.1.1 Writing full dataset

2. Spark

2.1. Writing CSV data

2.1.1 Writing to dataset

2.2. Writing Parquet data

2.2.1 Writing to dataset

from amorphicutils.pyspark.write import Write
from amorphicutils.pyspark.infra.gluespark. import GlueSpark

lz_bucket=""
domain=""
dataset=""
user=""

glue_spark = GlueSpark()
glue_context = glue_spark.get_glue_context()

csv_writer = Write(bucket_name=lz_bucket,
spark=glue_context)
result = csv_writer.write_parquet(df, domain, dataset, user)

If the dataset is of reload type then pass full_reload=True as parameter to the write_parquet method