From f91ee02ff52c5b944820ebf9d2b97ef7187390fd Mon Sep 17 00:00:00 2001 From: Felix Hennig Date: Thu, 10 Aug 2023 09:48:45 +0200 Subject: [PATCH] Added required external components --- .../druid/pages/required-external-components.adoc | 9 +++++++++ docs/modules/druid/pages/usage-guide/deep-storage.adoc | 6 ++++-- 2 files changed, 13 insertions(+), 2 deletions(-) create mode 100644 docs/modules/druid/pages/required-external-components.adoc diff --git a/docs/modules/druid/pages/required-external-components.adoc b/docs/modules/druid/pages/required-external-components.adoc new file mode 100644 index 00000000..2d9bcf38 --- /dev/null +++ b/docs/modules/druid/pages/required-external-components.adoc @@ -0,0 +1,9 @@ +# Required external components + +Druid uses an SQL database to store metadata. Consult the https://druid.apache.org/docs/latest/dependencies/metadata-storage.html#available-metadata-stores[Druid documentation] for a list of supported databases and setup instructions. + +## Feature specific: S3 and cloud deep storage + +https://druid.apache.org/docs/latest/dependencies/deep-storage.html[Deep storage] is where segments are stored. Druid offers multiple storage backends. For the local storage there are no prerequisites. HDFS deep storage can be set up with the xref:hdfs:index.adoc[Stackable Operator for Apache HDFS]. For S3 deep storage or the Google Cloud and Azure storage backends, you need to set up the storage. + +Read the xref:usage-guide/deep-storage.adoc[deep storage usage guide] to learn more about configuring Druid deep storage. diff --git a/docs/modules/druid/pages/usage-guide/deep-storage.adoc b/docs/modules/druid/pages/usage-guide/deep-storage.adoc index 25dd7100..5de6a228 100644 --- a/docs/modules/druid/pages/usage-guide/deep-storage.adoc +++ b/docs/modules/druid/pages/usage-guide/deep-storage.adoc @@ -1,8 +1,10 @@ = Deep storage configuration +https://druid.apache.org/docs/latest/dependencies/deep-storage.html[Deep Storage] is where Druid stores data segments. For a Kubernetes environment, either the HDFS or S3 backend is recommended. + == [[hdfs]]HDFS -Druid can use HDFS as a backend for deep storage: +Druid can use HDFS as a backend for deep storage, which requires having an HDFS instance running. You can use the xref:hdfs:index.adoc[Stackable Operator for Apache HDFS] to run HDFS. Configure the HDFS deep storage backend in your Druid cluster this way: [source,yaml] ---- @@ -80,4 +82,4 @@ include::partial$s3-note.adoc[] === S3 Credentials -include::partial$s3-credentials.adoc[] \ No newline at end of file +include::partial$s3-credentials.adoc[]