Copy Data from Snowflake to AWS
To copy data from Snowflake into AWS S3 without creating storage integration, you can use a named external stage. This way, you define the credentials and S3 bucket details within the stage itself rather than creating a storage integration. Here's how you can do it:
Create the external stage
Copy data into the S3 bucket
Hereβs the process step-by-step:
Create the external stage:
Replace your-bucket-name
, your-path
, your-aws-access-key-id
, and your-aws-secret-access-key
with your actual S3 bucket name, path, AWS access key ID, and secret access key, respectively.
CREATE OR REPLACE STAGE my_s3_stage
URL = 's3://your-bucket-name/your-path/'
CREDENTIALS = (
AWS_KEY_ID = 'your-aws-access-key-id'
AWS_SECRET_KEY = 'your-aws-secret-access-key'
);
Copy data into the S3 bucket:
Replace your-file-name
with the desired file name in the S3 bucket and your_table
with the name of the table you want to export.
COPY INTO @my_s3_stage/your-file-name
FROM your_table
FILE_FORMAT = (TYPE = CSV);
This way, you donβt need to create a storage integration, and you can still copy data from Snowflake to S3.
Last updated
Was this helpful?