Skip to main content

How to create and run a Spark ETL job?

headerImage

info

Steps to be followed

  • Step1: Click here to transfer data from a Postgres DB to Amorphic datasets.
    • This step creates two datasets retail_sales_raw_<your_userid>, retail_store_info_<your_userid> that are needed as inputs to the Spark job (Step3).
    • If you've already ran this job earlier, skip to step #2.
  • Step2: Click here to create the output datasets needed for the Spark Job.
  • Step3: Click here to create and run the Spark Job.


tip

Now, proceed to run a drag-and-drop Morph job.