Delta Table

Connect delta tables to your preprocessing pipeline, and batch process all your records using unstructured-ingest to store structured outputs locally on your filesystem.

First you’ll need to install the delta table dependencies as shown here.

pip install "unstructured[delta-table]"

Run Locally

#!/usr/bin/env bash

unstructured-ingest \
  delta-table \
  --table-uri s3://utic-dev-tech-fixtures/sample-delta-lake-data/deltatable/ \
  --output-dir delta-table-example \
  --storage_options "AWS_REGION=us-east-2,AWS_ACCESS_KEY_ID=$AWS_ACCESS_KEY_ID,AWS_SECRET_ACCESS_KEY=$AWS_SECRET_ACCESS_KEY" \
  --verbose

Run via the API

You can also use upstream connectors with the unstructured API. For this you’ll need to use the --partition-by-api flag and pass in your API key with --api-key.

#!/usr/bin/env bash

unstructured-ingest \
  delta-table \
  --table-uri s3://utic-dev-tech-fixtures/sample-delta-lake-data/deltatable/ \
  --output-dir delta-table-example \
  --storage_options "AWS_REGION=us-east-2,AWS_ACCESS_KEY_ID=$AWS_ACCESS_KEY_ID,AWS_SECRET_ACCESS_KEY=$AWS_SECRET_ACCESS_KEY" \
  --verbose \
  --partition-by-api \
  --api-key "<UNSTRUCTURED-API-KEY>"

Additionally, you will need to pass the --partition-endpoint if you’re running the API locally. You can find more information about the unstructured API here.

For a full list of the options the CLI accepts check unstructured-ingest delta-table --help.

NOTE: Keep in mind that you will need to have all the appropriate extras and dependencies for the file types of the documents contained in your data storage platform if you’re running this locally. You can find more information about this in the installation guide.