Sha256: d664bb2b822c4f360b9e0b66f4d2c37615a20d67b2269918f9cd4c12c76ddebe
Contents?: true
Size: 1.58 KB
Versions: 3
Compression:
Stored size: 1.58 KB
Contents
## case study It is a guide to replicate single mysql table to elasticsearch. ## configuration ``` <source> type mysql_replicator # Set connection settings for replicate source. host localhost username your_mysql_user password your_mysql_password database myweb # Set replicate query configuration. query SELECT id, text, updated_at from search_test; primary_key id # specify unique key (default: id) interval 10s # execute query interval (default: 1m) # Enable detect deletion event not only insert/update events. (default: yes) # It is useful to use `enable_delete no` that keep following recently updated record with this query. # `SELECT * FROM search_test WHERE DATE_ADD(updated_at, INTERVAL 5 MINUTE) > NOW();` enable_delete yes # Format output tag for each events. Placeholders usage as described below. tag replicator.myweb.search_test.${event}.${primary_key} # ${event} : the variation of row event type by insert/update/delete. # ${primary_key} : the value of `replicator_manager.settings.primary_key` in manager table. </source> <match replicator.**> type mysql_replicator_elasticsearch # Set Elasticsearch connection. host localhost port 9200 # Set Elasticsearch index, type, and unique id (primary_key) from tag. tag_format (?<core_name>[^\.]+)\.(?<event>[^\.]+)\.(?<primary_key>[^\.]+)$ # Set frequency of sending bulk request to Elasticsearch node. flush_interval 5s # Set maximum retry interval (required fluentd >= 0.10.41) max_retry_wait 1800 # Queued chunks are flushed at shutdown process. flush_at_shutdown yes </match> ```
Version data entries
3 entries across 3 versions & 1 rubygems