From 8ed75adf1ecfdff7fa1929b3eb90311c6016a523 Mon Sep 17 00:00:00 2001 From: dat-a-man <98139823+dat-a-man@users.noreply.github.com> Date: Sat, 25 May 2024 03:42:59 +0000 Subject: [PATCH] updated installation command in destination docs and a few others --- docs/website/docs/dlt-ecosystem/destinations/athena.md | 4 ++-- docs/website/docs/dlt-ecosystem/destinations/bigquery.md | 2 +- docs/website/docs/dlt-ecosystem/destinations/clickhouse.md | 4 ++-- docs/website/docs/dlt-ecosystem/destinations/databricks.md | 2 +- docs/website/docs/dlt-ecosystem/destinations/dremio.md | 4 ++-- docs/website/docs/dlt-ecosystem/destinations/duckdb.md | 2 +- docs/website/docs/dlt-ecosystem/destinations/filesystem.md | 6 +++--- docs/website/docs/dlt-ecosystem/destinations/motherduck.md | 2 +- docs/website/docs/dlt-ecosystem/destinations/mssql.md | 4 ++-- docs/website/docs/dlt-ecosystem/destinations/postgres.md | 2 +- docs/website/docs/dlt-ecosystem/destinations/qdrant.md | 2 +- docs/website/docs/dlt-ecosystem/destinations/redshift.md | 4 ++-- docs/website/docs/dlt-ecosystem/destinations/synapse.md | 2 +- docs/website/docs/dlt-ecosystem/destinations/weaviate.md | 2 +- docs/website/docs/dlt-ecosystem/file-formats/parquet.md | 2 +- .../data-enrichments/currency_conversion_data_enrichment.md | 2 +- .../data-enrichments/url-parser-data-enrichment.md | 2 +- .../data-enrichments/user_agent_device_data_enrichment.md | 2 +- 18 files changed, 25 insertions(+), 25 deletions(-) diff --git a/docs/website/docs/dlt-ecosystem/destinations/athena.md b/docs/website/docs/dlt-ecosystem/destinations/athena.md index 76491578fe..7c907664d3 100644 --- a/docs/website/docs/dlt-ecosystem/destinations/athena.md +++ b/docs/website/docs/dlt-ecosystem/destinations/athena.md @@ -11,7 +11,7 @@ The Athena destination stores data as Parquet files in S3 buckets and creates [e ## Install dlt with Athena **To install the dlt library with Athena dependencies:** ```sh -pip install dlt[athena] +pip install "dlt[athena]" ``` ## Setup Guide @@ -30,7 +30,7 @@ First, install dependencies by running: ```sh pip install -r requirements.txt ``` -or with `pip install dlt[athena]`, which will install `s3fs`, `pyarrow`, `pyathena`, and `botocore` packages. +or with `pip install "dlt[athena]"`, which will install `s3fs`, `pyarrow`, `pyathena`, and `botocore` packages. :::caution diff --git a/docs/website/docs/dlt-ecosystem/destinations/bigquery.md b/docs/website/docs/dlt-ecosystem/destinations/bigquery.md index 54d5abae6d..4f99901e37 100644 --- a/docs/website/docs/dlt-ecosystem/destinations/bigquery.md +++ b/docs/website/docs/dlt-ecosystem/destinations/bigquery.md @@ -11,7 +11,7 @@ keywords: [bigquery, destination, data warehouse] **To install the dlt library with BigQuery dependencies:** ```sh -pip install dlt[bigquery] +pip install "dlt[bigquery]" ``` ## Setup Guide diff --git a/docs/website/docs/dlt-ecosystem/destinations/clickhouse.md b/docs/website/docs/dlt-ecosystem/destinations/clickhouse.md index ea187e54eb..58551751c5 100644 --- a/docs/website/docs/dlt-ecosystem/destinations/clickhouse.md +++ b/docs/website/docs/dlt-ecosystem/destinations/clickhouse.md @@ -11,7 +11,7 @@ keywords: [ clickhouse, destination, data warehouse ] **To install the DLT library with ClickHouse dependencies:** ```sh -pip install dlt[clickhouse] +pip install "dlt[clickhouse]" ``` ## Setup Guide @@ -33,7 +33,7 @@ requirements file by executing it as follows: pip install -r requirements.txt ``` -or with `pip install dlt[clickhouse]`, which installs the `dlt` library and the necessary dependencies for working with ClickHouse as a destination. +or with `pip install "dlt[clickhouse]"`, which installs the `dlt` library and the necessary dependencies for working with ClickHouse as a destination. ### 2. Setup ClickHouse database diff --git a/docs/website/docs/dlt-ecosystem/destinations/databricks.md b/docs/website/docs/dlt-ecosystem/destinations/databricks.md index b601809935..6cd5767dcb 100644 --- a/docs/website/docs/dlt-ecosystem/destinations/databricks.md +++ b/docs/website/docs/dlt-ecosystem/destinations/databricks.md @@ -12,7 +12,7 @@ keywords: [Databricks, destination, data warehouse] ## Install dlt with Databricks **To install the dlt library with Databricks dependencies:** ```sh -pip install dlt[databricks] +pip install "dlt[databricks]" ``` ## Set up your Databricks workspace diff --git a/docs/website/docs/dlt-ecosystem/destinations/dremio.md b/docs/website/docs/dlt-ecosystem/destinations/dremio.md index 0be01e8e32..546f470938 100644 --- a/docs/website/docs/dlt-ecosystem/destinations/dremio.md +++ b/docs/website/docs/dlt-ecosystem/destinations/dremio.md @@ -9,7 +9,7 @@ keywords: [dremio, iceberg, aws, glue catalog] ## Install dlt with Dremio **To install the dlt library with Dremio and s3 dependencies:** ```sh -pip install dlt[dremio,s3] +pip install "dlt[dremio,s3]" ``` ## Setup Guide @@ -28,7 +28,7 @@ First install dependencies by running: ```sh pip install -r requirements.txt ``` -or with `pip install dlt[dremio,s3]` which will install `s3fs`, `pyarrow`, and `botocore` packages. +or with `pip install "dlt[dremio,s3]"` which will install `s3fs`, `pyarrow`, and `botocore` packages. To edit the `dlt` credentials file with your secret info, open `.dlt/secrets.toml`. You will need to provide a `bucket_url` which holds the uploaded parquet files. diff --git a/docs/website/docs/dlt-ecosystem/destinations/duckdb.md b/docs/website/docs/dlt-ecosystem/destinations/duckdb.md index 22c5fd1df9..c2f6786f8d 100644 --- a/docs/website/docs/dlt-ecosystem/destinations/duckdb.md +++ b/docs/website/docs/dlt-ecosystem/destinations/duckdb.md @@ -9,7 +9,7 @@ keywords: [duckdb, destination, data warehouse] ## Install dlt with DuckDB **To install the dlt library with DuckDB dependencies, run:** ```sh -pip install dlt[duckdb] +pip install "dlt[duckdb]" ``` ## Setup Guide diff --git a/docs/website/docs/dlt-ecosystem/destinations/filesystem.md b/docs/website/docs/dlt-ecosystem/destinations/filesystem.md index 0d719b4cfa..4c62e172d8 100644 --- a/docs/website/docs/dlt-ecosystem/destinations/filesystem.md +++ b/docs/website/docs/dlt-ecosystem/destinations/filesystem.md @@ -6,7 +6,7 @@ The Filesystem destination stores data in remote file systems and bucket storage ## Install dlt with filesystem **To install the dlt library with filesystem dependencies:** ```sh -pip install dlt[filesystem] +pip install "dlt[filesystem]" ``` This installs `s3fs` and `botocore` packages. @@ -125,7 +125,7 @@ client_kwargs = '{"verify": "public.crt"}' ``` #### Google Storage -Run `pip install dlt[gs]` which will install the `gcfs` package. +Run `pip install "dlt[gs]"` which will install the `gcfs` package. To edit the `dlt` credentials file with your secret info, open `.dlt/secrets.toml`. You'll see AWS credentials by default. @@ -148,7 +148,7 @@ if you have default google cloud credentials in your environment (i.e. on cloud Use **Cloud Storage** admin to create a new bucket. Then assign the **Storage Object Admin** role to your service account. #### Azure Blob Storage -Run `pip install dlt[az]` which will install the `adlfs` package to interface with Azure Blob Storage. +Run `pip install "dlt[az]"` which will install the `adlfs` package to interface with Azure Blob Storage. Edit the credentials in `.dlt/secrets.toml`, you'll see AWS credentials by default replace them with your Azure credentials: diff --git a/docs/website/docs/dlt-ecosystem/destinations/motherduck.md b/docs/website/docs/dlt-ecosystem/destinations/motherduck.md index b053d29ac1..9d8c8d260b 100644 --- a/docs/website/docs/dlt-ecosystem/destinations/motherduck.md +++ b/docs/website/docs/dlt-ecosystem/destinations/motherduck.md @@ -10,7 +10,7 @@ keywords: [MotherDuck, duckdb, destination, data warehouse] ## Install dlt with MotherDuck **To install the dlt library with MotherDuck dependencies:** ```sh -pip install dlt[motherduck] +pip install "dlt[motherduck]" ``` :::tip diff --git a/docs/website/docs/dlt-ecosystem/destinations/mssql.md b/docs/website/docs/dlt-ecosystem/destinations/mssql.md index 3e5b209aaa..1cba484f10 100644 --- a/docs/website/docs/dlt-ecosystem/destinations/mssql.md +++ b/docs/website/docs/dlt-ecosystem/destinations/mssql.md @@ -9,7 +9,7 @@ keywords: [mssql, sqlserver, destination, data warehouse] ## Install dlt with MS SQL **To install the dlt library with MS SQL dependencies, use:** ```sh -pip install dlt[mssql] +pip install "dlt[mssql]" ``` ## Setup guide @@ -38,7 +38,7 @@ pip install -r requirements.txt ``` or run: ```sh -pip install dlt[mssql] +pip install "dlt[mssql]" ``` This will install `dlt` with the `mssql` extra, which contains all the dependencies required by the SQL server client. diff --git a/docs/website/docs/dlt-ecosystem/destinations/postgres.md b/docs/website/docs/dlt-ecosystem/destinations/postgres.md index 5126272e37..ae504728c3 100644 --- a/docs/website/docs/dlt-ecosystem/destinations/postgres.md +++ b/docs/website/docs/dlt-ecosystem/destinations/postgres.md @@ -9,7 +9,7 @@ keywords: [postgres, destination, data warehouse] ## Install dlt with PostgreSQL **To install the dlt library with PostgreSQL dependencies, run:** ```sh -pip install dlt[postgres] +pip install "dlt[postgres]" ``` ## Setup Guide diff --git a/docs/website/docs/dlt-ecosystem/destinations/qdrant.md b/docs/website/docs/dlt-ecosystem/destinations/qdrant.md index 1b560ad6fe..9f19007227 100644 --- a/docs/website/docs/dlt-ecosystem/destinations/qdrant.md +++ b/docs/website/docs/dlt-ecosystem/destinations/qdrant.md @@ -14,7 +14,7 @@ This destination helps you load data into Qdrant from [dlt resources](../../gene 1. To use Qdrant as a destination, make sure `dlt` is installed with the `qdrant` extra: ```sh -pip install dlt[qdrant] +pip install "dlt[qdrant]" ``` 2. Next, configure the destination in the dlt secrets file. The file is located at `~/.dlt/secrets.toml` by default. Add the following section to the secrets file: diff --git a/docs/website/docs/dlt-ecosystem/destinations/redshift.md b/docs/website/docs/dlt-ecosystem/destinations/redshift.md index 349698d201..7e0679ec6b 100644 --- a/docs/website/docs/dlt-ecosystem/destinations/redshift.md +++ b/docs/website/docs/dlt-ecosystem/destinations/redshift.md @@ -9,7 +9,7 @@ keywords: [redshift, destination, data warehouse] ## Install dlt with Redshift **To install the dlt library with Redshift dependencies:** ```sh -pip install dlt[redshift] +pip install "dlt[redshift]" ``` ## Setup Guide @@ -26,7 +26,7 @@ The above command generates several files and directories, including `.dlt/secre ```sh pip install -r requirements.txt ``` -or with `pip install dlt[redshift]`, which installs the `dlt` library and the necessary dependencies for working with Amazon Redshift as a destination. +or with `pip install "dlt[redshift]"`, which installs the `dlt` library and the necessary dependencies for working with Amazon Redshift as a destination. ### 2. Setup Redshift cluster To load data into Redshift, you need to create a Redshift cluster and enable access to your IP address through the VPC inbound rules associated with the cluster. While we recommend asking our GPT-4 assistant for details, we have provided a general outline of the process below: diff --git a/docs/website/docs/dlt-ecosystem/destinations/synapse.md b/docs/website/docs/dlt-ecosystem/destinations/synapse.md index f1c43b4d54..2e936f193e 100644 --- a/docs/website/docs/dlt-ecosystem/destinations/synapse.md +++ b/docs/website/docs/dlt-ecosystem/destinations/synapse.md @@ -9,7 +9,7 @@ keywords: [synapse, destination, data warehouse] ## Install dlt with Synapse **To install the dlt library with Synapse dependencies:** ```sh -pip install dlt[synapse] +pip install "dlt[synapse]" ``` ## Setup guide diff --git a/docs/website/docs/dlt-ecosystem/destinations/weaviate.md b/docs/website/docs/dlt-ecosystem/destinations/weaviate.md index 1272b16c86..11d1276ceb 100644 --- a/docs/website/docs/dlt-ecosystem/destinations/weaviate.md +++ b/docs/website/docs/dlt-ecosystem/destinations/weaviate.md @@ -14,7 +14,7 @@ This destination helps you load data into Weaviate from [dlt resources](../../ge 1. To use Weaviate as a destination, make sure dlt is installed with the 'weaviate' extra: ```sh -pip install dlt[weaviate] +pip install "dlt[weaviate]" ``` 2. Next, configure the destination in the dlt secrets file. The file is located at `~/.dlt/secrets.toml` by default. Add the following section to the secrets file: diff --git a/docs/website/docs/dlt-ecosystem/file-formats/parquet.md b/docs/website/docs/dlt-ecosystem/file-formats/parquet.md index 8944b7d5fa..414eaf2cb8 100644 --- a/docs/website/docs/dlt-ecosystem/file-formats/parquet.md +++ b/docs/website/docs/dlt-ecosystem/file-formats/parquet.md @@ -11,7 +11,7 @@ keywords: [parquet, file formats] To use this format, you need a `pyarrow` package. You can get this package as a `dlt` extra as well: ```sh -pip install dlt[parquet] +pip install "dlt[parquet]" ``` ## Supported Destinations diff --git a/docs/website/docs/general-usage/data-enrichments/currency_conversion_data_enrichment.md b/docs/website/docs/general-usage/data-enrichments/currency_conversion_data_enrichment.md index f8bd179422..82297420ed 100644 --- a/docs/website/docs/general-usage/data-enrichments/currency_conversion_data_enrichment.md +++ b/docs/website/docs/general-usage/data-enrichments/currency_conversion_data_enrichment.md @@ -247,7 +247,7 @@ API token. [destination](../../dlt-ecosystem/destinations/), For example, duckdb: ```sh - pip install dlt[duckdb] + pip install "dlt[duckdb]" ``` 1. Run the pipeline with the following command: diff --git a/docs/website/docs/general-usage/data-enrichments/url-parser-data-enrichment.md b/docs/website/docs/general-usage/data-enrichments/url-parser-data-enrichment.md index ab71d3d1d0..f2cd4a1065 100644 --- a/docs/website/docs/general-usage/data-enrichments/url-parser-data-enrichment.md +++ b/docs/website/docs/general-usage/data-enrichments/url-parser-data-enrichment.md @@ -231,7 +231,7 @@ need to register to use this service neither get an API key. [destination](https://dlthub.com/docs/dlt-ecosystem/destinations/), For example, duckdb: ```sh - pip install dlt[duckdb] + pip install "dlt[duckdb]" ``` 1. Run the pipeline with the following command: diff --git a/docs/website/docs/general-usage/data-enrichments/user_agent_device_data_enrichment.md b/docs/website/docs/general-usage/data-enrichments/user_agent_device_data_enrichment.md index 3aadb2f982..2448d31a06 100644 --- a/docs/website/docs/general-usage/data-enrichments/user_agent_device_data_enrichment.md +++ b/docs/website/docs/general-usage/data-enrichments/user_agent_device_data_enrichment.md @@ -284,7 +284,7 @@ The first step is to register on [SerpAPI](https://serpapi.com/) and obtain the [destination](https://dlthub.com/docs/dlt-ecosystem/destinations/), For example, duckdb: ```sh - pip install dlt[duckdb] + pip install "dlt[duckdb]" ``` 1. Run the pipeline with the following command: