# Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY # KIND, either express or implied. See the License for the # specific language governing permissions and limitations # under the License. LANG=C.UTF-8 LANGUAGE=C.UTF-8 LC_ALL=C.UTF-8 # JAVA OPTS # -XX:HeapDumpPath set in container based on the service and intance DRUID_COMMON_JAVA_OPTS=-server -Duser.timezone=UTC -Dfile.encoding=UTF-8 -XX:+UseG1GC -XX:+ExitOnOutOfMemoryError -XX:+HeapDumpOnOutOfMemoryError DRUID_SERVICE_JAVA_OPTS= # Debugging. Enabled by default. Always uses port 8000 in the container. # May to unique ports onto the host in Docker Compose. DEBUG_PORT= DEBUG_OPTS=-agentlib:jdwp=transport=dt_socket,server=y,suspend=n,address=8000 # Extra classpath. Standard classpath includes Druid libs, Hadoop config, Druid config. DRUID_CLASSPATH= # Number of an instance when there are two or more of the same type. DRUID_INSTANCE= # Druid configs # Mostly adapted from the micro-quickstart/_common file. # Hostname # druid.host is set on each host by the launch script # Extensions specified in the load list will be loaded by Druid at runtime. # The extension jars must be installed as part of Druid, or via the image # build script. # # The launch script creates druid_extensions_loadList by combining two test-specific # variables: druid_standard_loadList defined here, and druid_test_loadList, defined # in a docker-compose.yaml file, for any test-specific extensions. # See compose.md for more details. druid_standard_loadList=mysql-metadata-storage,druid-it-tools,druid-lookups-cached-global,druid-histogram,druid-datasketches,druid-parquet-extensions,druid-avro-extensions,druid-protobuf-extensions,druid-orc-extensions,druid-kafka-indexing-service,druid-s3-extensions,druid-multi-stage-query,druid-catalog # Location of Hadoop dependencies provided at runtime in the shared directory. druid_extensions_hadoopDependenciesDir=/shared/hadoop-dependencies # Logging druid_startup_logging_logProperties=true # Zookeeper # Name of the ZK container, mapped to a host name druid_zk_service_host=zookeeper druid_zk_paths_base=/druid # Metadata storage # For MySQL (MySQL JDBC installed in $DRUID_HOME/lib) druid_metadata_storage_type=mysql druid_metadata_storage_connector_connectURI=jdbc:mysql://metadata/druid druid_metadata_storage_connector_user=druid druid_metadata_storage_connector_password=diurd # Deep storage druid_storage_type=local druid_storage_storageDirectory=/shared/druid/storage # Indexing service logs druid_indexer_logs_type=file druid_indexer_logs_directory=/shared/tasklogs # Service discovery druid_selectors_indexing_serviceName=druid/overlord druid_selectors_coordinator_serviceName=druid/coordinator # Monitoring druid_monitoring_monitors=["org.apache.druid.java.util.metrics.JvmMonitor"] druid_emitter=noop druid_emitter_logging_logLevel=info # Storage type of double columns druid_indexing_doubleStorage=double # SQL druid_sql_enable=true druid_sql_planner_authorizeSystemTablesDirectly=true # Lookups druid_lookup_numLookupLoadingThreads=1 # Test-specific druid_server_http_numThreads=20 # Allow OPTIONS method for ITBasicAuthConfigurationTest.testSystemSchemaAccess druid_auth_basic_common_maxSyncRetries=20 druid_request_logging_type=slf4j # TODO: Can these 4 move to the Coordinator config? druid_coordinator_kill_supervisor_on=true druid_coordinator_kill_supervisor_period=PT10S druid_coordinator_kill_supervisor_durationToRetain=PT0M druid_coordinator_period_metadataStoreManagementPeriod=PT10S # TODO: Can the following be moved to Overlord? Or removed? # Testing the legacy config from https://github.com/apache/druid/pull/10267 # Can remove this when the flag is no longer needed druid_indexer_task_ignoreTimestampSpecForDruidInputSource=true # TODO: Pass this from the test (AzureDeepStorage)