![]() There are three methods of authenticating this connection: Redshift also connects to S3 during COPY and UNLOAD queries. save () // Write back to a table using IAM Role based authentication df. load () // After you have applied transformations to the data, you can use // the data source API to write the data back to another table // Write back to a table df. option ( "forward_spark_s3_credentials", True ). option ( "query", "select x, count(*) group by x" ). load () // Read data from a query val df = spark. Read data from a table val df = spark. ![]() save () ) # Write back to a table using IAM Role based authentication ( df. load () ) # After you have applied transformations to the data, you can use # the data source API to write the data back to another table # Write back to a table ( df. load () ) # Read data from a query df = ( spark.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |