-
Notifications
You must be signed in to change notification settings - Fork 13
/
Copy pathelasticsearch.yml.example
81 lines (81 loc) · 2.87 KB
/
elasticsearch.yml.example
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
## ================== Crawler Configuration - Elasticsearch ====================
#
## Elasticsearch connection settings. These can be defined individually
## for a crawl in the crawl config file.
#
## NOTE: Most Crawler configurations comes with reasonable defaults.
## Before adjusting the configuration, make sure you understand what you
## are trying to accomplish and the consequences.
#
## ------------------------------- Elasticsearch -------------------------------
#
## The host of the Elasticsearch deployment, excluding the port.
## If your Elasticsearch is running on Docker and is on the same machine as
## Crawler, this value should be `http://host.docker.internal`
#elasticsearch.host: http://localhost
#
#
## The port of the Elasticsearch deployment host.
#elasticsearch.port: 9200
#
#
## The encoded API key for Elasticsearch connection.
## Using `api_key` is recommended instead of `username`/`password`.
## Ensure this API key has read and write access to the configured
## `output_index` in the Crawler config
#elasticsearch.api_key: 1234
#
#
## The username for the Elasticsearch connection.
## Using `username` requires `password` to also be configured.
## However, `elasticsearch.api_key` is the recommended configuration choice.
#elasticsearch.username: elastic
#
#
## The password for the Elasticsearch connection.
## Using `password` requires `username` to also be configured.
## However, `elasticsearch.api_key` is the recommended configuration choice.
#elasticsearch.password: changeme
#
#
## The SHA256 CA cert fingerprint used to verify SSL connection to Elasticsearch.
## SSL usage is configured by the presence of `https` in `elasticsearch.host`
#elasticsearch.ca_fingerprint: null
#
#
## ------------------------------- Bulk API -----------------------------------
#
## The max size of the bulk queue
#elasticsearch.bulk_api.max_items: 10
#
#
## The max size in bytes of the bulk queue.
## When it's reached, the Crawler performs a bulk index request to
## Elasticsearch, and the queue is flushed
#elasticsearch.bulk_api.max_size_bytes: 1_048_576
#
#
## ------------------------------- Pipelines ----------------------------------
#
#
## The name of the ingest pipeline
## If pipelines are enabled and this value is `null`,
## the pipeline `ent-search-generic-ingestion` will be used
#elasticsearch.pipeline: ent-search-generic-ingestion
#
#
## Whether or not pipelines are enabled
#elasticsearch.pipeline_enabled: true
#
#
## Enable for the pipeline to reduce whitespace on indexed docs
#elasticsearch.pipeline_params._reduce_whitespace: true
#
#
## Enable for the pipeline to run ML inference on indexed docs
#elasticsearch.pipeline_params._run_ml_inference: true
#
#
## Enable for the pipeline to extract binary content from
## the `_attachment` field of an indexed doc
#elasticsearch.pipeline_params._extract_binary_content: true