Compare commits
56 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
3dc7627782 | ||
|
|
9235c48c88 | ||
|
|
53e665bbb6 | ||
|
|
fa2d226fbf | ||
|
|
da5a3d8be3 | ||
|
|
b10462dacd | ||
|
|
61c7a1307e | ||
|
|
b5419813ba | ||
|
|
ded1106b13 | ||
|
|
2b27f39088 | ||
|
|
7b337a8b91 | ||
|
|
927e532b2a | ||
|
|
26a32a3f08 | ||
|
|
6bb84b165f | ||
|
|
4e0292d222 | ||
|
|
909cae01b3 | ||
|
|
6f2bd2ab3e | ||
|
|
c5aeae1b02 | ||
|
|
a7d5a2e623 | ||
|
|
3a64a22ac4 | ||
|
|
c4b62769b9 | ||
|
|
b9e5f64d40 | ||
|
|
c0e358aafb | ||
|
|
442ddf16eb | ||
|
|
4e7985dafd | ||
|
|
ae51d77f05 | ||
|
|
529c98aadb | ||
|
|
bfcd9bf69a | ||
|
|
af55fde54a | ||
|
|
9e05a01dff | ||
|
|
064647607e | ||
|
|
1ece7f9abc | ||
|
|
38b7096419 | ||
|
|
eef7473a0b | ||
|
|
7a1da5b7cd | ||
|
|
e56176bbea | ||
|
|
49e751a9f8 | ||
|
|
9804850714 | ||
|
|
a6c669cc52 | ||
|
|
e615829310 | ||
|
|
362e9ad0a0 | ||
|
|
4994cd810b | ||
|
|
1487b41b3e | ||
|
|
dd29d16a31 | ||
|
|
ebe5596469 | ||
|
|
275cd6fc2f | ||
|
|
7da6317083 | ||
|
|
470e6309b5 | ||
|
|
a7dde52b7b | ||
|
|
53c0f5761d | ||
|
|
4287c01037 | ||
|
|
4f93fa7224 | ||
|
|
b52a7358ff | ||
|
|
beb8f72560 | ||
|
|
6b6973d6ee | ||
|
|
c790a645f2 |
12
.gitignore
vendored
Normal file
12
.gitignore
vendored
Normal file
@@ -0,0 +1,12 @@
|
||||
*.gem
|
||||
Gemfile.lock
|
||||
Gemfile.bak
|
||||
.bundle
|
||||
.vagrant
|
||||
.mvn
|
||||
vendor
|
||||
lib/**/*.jar
|
||||
|
||||
.DS_Store
|
||||
*.swp
|
||||
*.log
|
||||
11
.travis.yml
Normal file
11
.travis.yml
Normal file
@@ -0,0 +1,11 @@
|
||||
sudo: required
|
||||
language: ruby
|
||||
cache: bundler
|
||||
rvm:
|
||||
- jruby
|
||||
before_script:
|
||||
- bundle exec rake vendor
|
||||
- bundle exec rake install_jars
|
||||
- ./scripts/travis-before_script.sh
|
||||
- source ./scripts/travis-variables.sh
|
||||
script: bundle exec rspec
|
||||
42
CHANGELOG.md
Normal file
42
CHANGELOG.md
Normal file
@@ -0,0 +1,42 @@
|
||||
# Change Log
|
||||
All notable changes to this project will be documented in this file, from 0.2.0.
|
||||
|
||||
## [0.3.0] - 2016-07-24
|
||||
- Brings tests from v5 branch, providing greater coverage
|
||||
- Removes bulk update support, due to inconsistent behaviour
|
||||
- Plugin now marked as threadsafe, meaning only 1 instance per-Logstash
|
||||
- Raises default max_pool_size to match the default number of workers (1 connection per worker)
|
||||
|
||||
## [0.2.10] - 2016-07-07
|
||||
- Support non-string entries in statement array
|
||||
- Adds backtrace to exception logging
|
||||
|
||||
## [0.2.9] - 2016-06-29
|
||||
- Fix NameError exception.
|
||||
- Moved log_jdbc_exception calls
|
||||
|
||||
## [0.2.7] - 2016-05-29
|
||||
- Backport retry exception logic from v5 branch
|
||||
- Backport improved timestamp compatibility from v5 branch
|
||||
|
||||
## [0.2.6] - 2016-05-02
|
||||
- Fix for exception infinite loop
|
||||
|
||||
## [0.2.5] - 2016-04-11
|
||||
### Added
|
||||
- Basic tests running against DerbyDB
|
||||
- Fix for converting Logstash::Timestamp to iso8601 from @hordijk
|
||||
|
||||
## [0.2.4] - 2016-04-07
|
||||
- Documentation fixes from @hordijk
|
||||
|
||||
## [0.2.3] - 2016-02-16
|
||||
- Bug fixes
|
||||
|
||||
## [0.2.2] - 2015-12-30
|
||||
- Bug fixes
|
||||
|
||||
## [0.2.1] - 2015-12-22
|
||||
- Support for connection pooling support added through HikariCP
|
||||
- Support for unsafe statement handling (allowing dynamic queries)
|
||||
- Altered exception handling to now count sequential flushes with exceptions thrown
|
||||
151
README.md
151
README.md
@@ -1,95 +1,84 @@
|
||||
# logstash-jdbc
|
||||
JDBC output plugin for Logstash.
|
||||
# logstash-output-jdbc
|
||||
|
||||
[](https://travis-ci.org/theangryangel/logstash-output-jdbc)
|
||||
|
||||
This plugin is provided as an external plugin and is not part of the Logstash project.
|
||||
|
||||
Currently untested with logstash 1.5+. Support is planned.
|
||||
This plugin allows you to output to SQL databases, using JDBC adapters.
|
||||
See below for tested adapters, and example configurations.
|
||||
|
||||
## Warning
|
||||
This has not yet been extensively tested with all JDBC drivers and may not yet work for you.
|
||||
|
||||
If you do find this works for a JDBC driver without an example, let me know and provide a small example configuration if you can.
|
||||
|
||||
This plugin does not bundle any JDBC jar files, and does expect them to be in a
|
||||
particular location. Please ensure you read the 4 installation lines below.
|
||||
|
||||
## Changelog
|
||||
See CHANGELOG.md
|
||||
|
||||
## Versions
|
||||
Released versions are available via rubygems, and typically tagged.
|
||||
|
||||
For development:
|
||||
- See master branch for logstash v5 (currently **development only**)
|
||||
- See v2.x branch for logstash v2
|
||||
- See v1.5 branch for logstash v1.5
|
||||
- See v1.4 branch for logstash 1.4
|
||||
|
||||
## Installation
|
||||
- Copy lib directory contents into your logstash installation.
|
||||
- Create the directory vendor/jar/jdbc in your logstash installation (`mkdir -p vendor/jar/jdbc/`)
|
||||
- Add JDBC jar files to vendor/jar/jdbc in your logstash installation
|
||||
- Configure
|
||||
- Run `bin/logstash-plugin install logstash-output-jdbc` in your logstash installation directory
|
||||
- Now either:
|
||||
- Use driver_jar_path in your configuraton to specify a path to your jar file
|
||||
- Or:
|
||||
- Create the directory vendor/jar/jdbc in your logstash installation (`mkdir -p vendor/jar/jdbc/`)
|
||||
- Add JDBC jar files to vendor/jar/jdbc in your logstash installation
|
||||
- And then configure (examples can be found in the examples directory)
|
||||
|
||||
## Configuration options
|
||||
* driver_class, string, JDBC driver class to load
|
||||
* connection_string, string, JDBC connection string
|
||||
* statement, array, an array of strings representing the SQL statement to run. Index 0 is the SQL statement that is prepared, all other array entries are passed in as parameters (in order). A parameter may either be a property of the event (i.e. "@timestamp", or "host") or a formatted string (i.e. "%{host} - %{message}" or "%{message}"). If a key is passed then it will be automatically converted as required for insertion into SQL. If it's a formatted string then it will be passed in verbatim.
|
||||
* flush_size, number, default = 1000, number of entries to buffer before sending to SQL
|
||||
* idle_flush_time, number, default = 1, number of idle seconds before sending data to SQL, even if the flush_size has not been reached. If you modify this value you should also consider altering max_repeat_exceptions_time
|
||||
* max_repeat_exceptions, number, default = 5, number of times the same exception can repeat before we stop logstash. Set to a value less than 1 if you never want it to stop
|
||||
* max_repeat_exceptions_time, number, default = 30, maxium number of seconds between exceptions before they're considered "different" exceptions. If you modify idle_flush_time you should consider this value
|
||||
|
||||
| Option | Type | Description | Required? | Default |
|
||||
| ------ | ---- | ----------- | --------- | ------- |
|
||||
| driver_class | String | Specify a driver class if autoloading fails | No | |
|
||||
| driver_auto_commit | Boolean | If the driver does not support auto commit, you should set this to false | No | True |
|
||||
| driver_jar_path | String | File path to jar file containing your JDBC driver. This is optional, and all JDBC jars may be placed in $LOGSTASH_HOME/vendor/jar/jdbc instead. | No | |
|
||||
| connection_string | String | JDBC connection URL | Yes | |
|
||||
| username | String | JDBC username - this is optional as it may be included in the connection string, for many drivers | No | |
|
||||
| password | String | JDBC password - this is optional as it may be included in the connection string, for many drivers | No | |
|
||||
| statement | Array | An array of strings representing the SQL statement to run. Index 0 is the SQL statement that is prepared, all other array entries are passed in as parameters (in order). A parameter may either be a property of the event (i.e. "@timestamp", or "host") or a formatted string (i.e. "%{host} - %{message}" or "%{message}"). If a key is passed then it will be automatically converted as required for insertion into SQL. If it's a formatted string then it will be passed in verbatim. | Yes | |
|
||||
| unsafe_statement | Boolean | If yes, the statement is evaluated for event fields - this allows you to use dynamic table names, etc. **This is highly dangerous** and you should **not** use this unless you are 100% sure that the field(s) you are passing in are 100% safe. Failure to do so will result in possible SQL injections. Please be aware that there is also a potential performance penalty as each event must be evaluated and inserted into SQL one at a time, where as when this is false multiple events are inserted at once. Example statement: [ "insert into %{table_name_field} (column) values(?)", "fieldname" ] | No | False |
|
||||
| max_pool_size | Number | Maximum number of connections to open to the SQL server at any 1 time. Default set to same as Logstash default number of workers | No | 24 |
|
||||
| connection_timeout | Number | Number of seconds before a SQL connection is closed | No | 2800 |
|
||||
| flush_size | Number | Maximum number of entries to buffer before sending to SQL - if this is reached before idle_flush_time | No | 1000 |
|
||||
| max_flush_exceptions | Number | Number of sequential flushes which cause an exception, before the set of events are discarded. Set to a value less than 1 if you never want it to stop. This should be carefully configured with respect to retry_initial_interval and retry_max_interval, if your SQL server is not highly available | No | 10 |
|
||||
| retry_initial_interval | Number | Number of seconds before the initial retry in the event of a failure. On each failure it will be doubled until it reaches retry_max_interval | No | 2 |
|
||||
| retry_max_interval | Number | Maximum number of seconds between each retry | No | 128 |
|
||||
| retry_sql_states | Array of strings | An array of custom SQL state codes you wish to retry until `max_flush_exceptions`. Useful if you're using a JDBC driver which returns retry-able, but non-standard SQL state codes in it's exceptions. | No | [] |
|
||||
|
||||
## Example configurations
|
||||
### SQLite3
|
||||
* Tested using https://bitbucket.org/xerial/sqlite-jdbc
|
||||
* SQLite setup - `echo "CREATE table log (host text, timestamp datetime, message text);" | sqlite3 test.db`
|
||||
```
|
||||
input
|
||||
{
|
||||
stdin { }
|
||||
}
|
||||
output {
|
||||
stdout { }
|
||||
Example logstash configurations, can now be found in the examples directory. Where possible we try to link every configuration with a tested jar.
|
||||
|
||||
jdbc {
|
||||
driver_class => 'org.sqlite.JDBC'
|
||||
connection_string => 'jdbc:sqlite:test.db'
|
||||
statement => [ "INSERT INTO log (host, timestamp, message) VALUES(?, ?, ?)", "host", "@timestamp", "message" ]
|
||||
}
|
||||
}
|
||||
```
|
||||
If you have a working sample configuration, for a DB thats not listed, pull requests are welcome.
|
||||
|
||||
### SQL Server
|
||||
* Tested using http://msdn.microsoft.com/en-gb/sqlserver/aa937724.aspx
|
||||
```
|
||||
input
|
||||
{
|
||||
stdin { }
|
||||
}
|
||||
output {
|
||||
jdbc {
|
||||
driver_class => 'com.microsoft.sqlserver.jdbc.SQLServerDriver'
|
||||
connection_string => "jdbc:sqlserver://server:1433;databaseName=databasename;user=username;password=password;autoReconnect=true;"
|
||||
statement => [ "INSERT INTO log (host, timestamp, message) VALUES(?, ?, ?)", "host", "@timestamp", "message" ]
|
||||
}
|
||||
}
|
||||
```
|
||||
## Development and Running tests
|
||||
For development tests are recommended to run inside a virtual machine (Vagrantfile is included in the repo), as it requires
|
||||
access to various database engines and could completely destroy any data in a live system.
|
||||
|
||||
### Postgres
|
||||
With thanks to [@roflmao](https://github.com/roflmao)
|
||||
```
|
||||
input
|
||||
{
|
||||
stdin { }
|
||||
}
|
||||
output {
|
||||
jdbc {
|
||||
driver_class => 'org.postgresql.Driver'
|
||||
connection_string => 'jdbc:postgresql://hostname:5432/database?user=username&password=password'
|
||||
statement => [ "INSERT INTO log (host, timestamp, message) VALUES(?, CAST (? AS timestamp), ?)", "host", "@timestamp", "message" ]
|
||||
}
|
||||
}
|
||||
```
|
||||
If you have vagrant available (this is temporary whilst I'm hacking on v5 support. I'll make this more streamlined later):
|
||||
- `vagrant up`
|
||||
- `vagrant ssh`
|
||||
- `cd /vagrant`
|
||||
- `gem install bundler`
|
||||
- `cd /vagrant && bundle install && bundle exec rake vendor && bundle exec rake install_jars`
|
||||
- `./scripts/travis-before_script.sh && source ./scripts/travis-variables.sh`
|
||||
- `bundle exec rspec`
|
||||
|
||||
### Oracle
|
||||
With thanks to [@josemazo](https://github.com/josemazo)
|
||||
* Tested with Express Edition 11g Release 2
|
||||
* Tested using http://www.oracle.com/technetwork/database/enterprise-edition/jdbc-112010-090769.html (ojdbc6.jar)
|
||||
```
|
||||
input
|
||||
{
|
||||
stdin { }
|
||||
}
|
||||
output {
|
||||
jdbc {
|
||||
driver_class => "oracle.jdbc.driver.OracleDriver"
|
||||
connection_string => "jdbc:oracle:thin:USER/PASS@HOST:PORT:SID"
|
||||
statement => [ "INSERT INTO log (host, timestamp, message) VALUES(?, CAST (? AS timestamp), ?)", "host", "@timestamp", "message" ]
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
/* vim: set ts=4 sw=4 tw=0 :*/
|
||||
## Releasing
|
||||
- Update Changelog
|
||||
- Bump version in gemspec
|
||||
- Commit
|
||||
- Create tag `git tag v<version-number-in-gemspec>`
|
||||
- `bundle exec rake install_jars`
|
||||
- `bundle exec rake pre_release_checks`
|
||||
- `gem build logstash-output-jdbc.gemspec`
|
||||
- `gem push`
|
||||
|
||||
11
Rakefile
Normal file
11
Rakefile
Normal file
@@ -0,0 +1,11 @@
|
||||
# encoding: utf-8
|
||||
require 'logstash/devutils/rake'
|
||||
require 'jars/installer'
|
||||
require 'rubygems'
|
||||
|
||||
desc 'Fetch any jars required for this plugin'
|
||||
task :install_jars do
|
||||
ENV['JARS_HOME'] = Dir.pwd + '/vendor/jar-dependencies/runtime-jars'
|
||||
ENV['JARS_VENDOR'] = 'false'
|
||||
Jars::Installer.new.vendor_jars!(false)
|
||||
end
|
||||
16
examples/apache-phoenix-hbase-sql.md
Normal file
16
examples/apache-phoenix-hbase-sql.md
Normal file
@@ -0,0 +1,16 @@
|
||||
# Example: Apache Phoenix (HBase SQL)
|
||||
* Tested with Ubuntu 14.04.03 / Logstash 2.1 / Apache Phoenix 4.6
|
||||
* <!> HBase and Zookeeper must be both accessible from logstash machine <!>
|
||||
```
|
||||
input
|
||||
{
|
||||
stdin { }
|
||||
}
|
||||
output {
|
||||
jdbc {
|
||||
connection_string => "jdbc:phoenix:ZOOKEEPER_HOSTNAME"
|
||||
statement => [ "UPSERT INTO EVENTS log (host, timestamp, message) VALUES(?, ?, ?)", "host", "@timestamp", "message" ]
|
||||
}
|
||||
|
||||
}
|
||||
```
|
||||
16
examples/mariadb.md
Normal file
16
examples/mariadb.md
Normal file
@@ -0,0 +1,16 @@
|
||||
# Example: MariaDB
|
||||
* Tested with Ubuntu 14.04.3 LTS, Server version: 10.1.9-MariaDB-1~trusty-log mariadb.org binary distribution
|
||||
* Tested using https://downloads.mariadb.com/enterprise/tqge-whfa/connectors/java/connector-java-1.3.2/mariadb-java-client-1.3.2.jar (mariadb-java-client-1.3.2.jar)
|
||||
```
|
||||
input
|
||||
{
|
||||
stdin { }
|
||||
}
|
||||
output {
|
||||
jdbc {
|
||||
connection_string => "jdbc:mariadb://HOSTNAME/DATABASE?user=USER&password=PASSWORD"
|
||||
statement => [ "INSERT INTO log (host, timestamp, message) VALUES(?, ?, ?)", "host", "@timestamp", "message" ]
|
||||
}
|
||||
|
||||
}
|
||||
```
|
||||
16
examples/mysql.md
Normal file
16
examples/mysql.md
Normal file
@@ -0,0 +1,16 @@
|
||||
# Example: Mysql
|
||||
With thanks to [@jMonsinjon](https://github.com/jMonsinjon)
|
||||
* Tested with Version 14.14 Distrib 5.5.43, for debian-linux-gnu (x86_64)
|
||||
* Tested using http://dev.mysql.com/downloads/file.php?id=457911 (mysql-connector-java-5.1.36-bin.jar)
|
||||
```
|
||||
input
|
||||
{
|
||||
stdin { }
|
||||
}
|
||||
output {
|
||||
jdbc {
|
||||
connection_string => "jdbc:mysql://HOSTNAME/DATABASE?user=USER&password=PASSWORD"
|
||||
statement => [ "INSERT INTO log (host, timestamp, message) VALUES(?, CAST (? AS timestamp), ?)", "host", "@timestamp", "message" ]
|
||||
}
|
||||
}
|
||||
```
|
||||
20
examples/odps.md
Normal file
20
examples/odps.md
Normal file
@@ -0,0 +1,20 @@
|
||||
# Example: ODPS
|
||||
With thanks to [@onesuper](https://github.com/onesuper)
|
||||
```
|
||||
input
|
||||
{
|
||||
stdin { }
|
||||
}
|
||||
output {
|
||||
jdbc {
|
||||
driver_class => "com.aliyun.odps.jdbc.OdpsDriver"
|
||||
driver_auto_commit => false
|
||||
connection_string => "jdbc:odps:http://service.odps.aliyun.com/api?project=meta_dev&loglevel=DEBUG"
|
||||
username => "abcd"
|
||||
password => "1234"
|
||||
max_pool_size => 5
|
||||
flush_size => 10
|
||||
statement => [ "INSERT INTO test_logstash VALUES(?, ?, ?);", "host", "@timestamp", "message" ]
|
||||
}
|
||||
}
|
||||
```
|
||||
16
examples/oracle.md
Normal file
16
examples/oracle.md
Normal file
@@ -0,0 +1,16 @@
|
||||
# Example: Oracle
|
||||
With thanks to [@josemazo](https://github.com/josemazo)
|
||||
* Tested with Express Edition 11g Release 2
|
||||
* Tested using http://www.oracle.com/technetwork/database/enterprise-edition/jdbc-112010-090769.html (ojdbc6.jar)
|
||||
```
|
||||
input
|
||||
{
|
||||
stdin { }
|
||||
}
|
||||
output {
|
||||
jdbc {
|
||||
connection_string => "jdbc:oracle:thin:USER/PASS@HOST:PORT:SID"
|
||||
statement => [ "INSERT INTO log (host, timestamp, message) VALUES(?, CAST (? AS timestamp), ?)", "host", "@timestamp", "message" ]
|
||||
}
|
||||
}
|
||||
```
|
||||
15
examples/postgres.md
Normal file
15
examples/postgres.md
Normal file
@@ -0,0 +1,15 @@
|
||||
# Example: Postgres
|
||||
With thanks to [@roflmao](https://github.com/roflmao)
|
||||
```
|
||||
input
|
||||
{
|
||||
stdin { }
|
||||
}
|
||||
output {
|
||||
jdbc {
|
||||
connection_string => 'jdbc:postgresql://hostname:5432/database?user=username&password=password'
|
||||
statement => [ "INSERT INTO log (host, timestamp, message) VALUES(?, CAST (? AS timestamp), ?)", "host", "@timestamp", "message" ]
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
14
examples/sql-server.md
Normal file
14
examples/sql-server.md
Normal file
@@ -0,0 +1,14 @@
|
||||
# Example: SQL Server
|
||||
* Tested using http://msdn.microsoft.com/en-gb/sqlserver/aa937724.aspx
|
||||
```
|
||||
input
|
||||
{
|
||||
stdin { }
|
||||
}
|
||||
output {
|
||||
jdbc {
|
||||
connection_string => "jdbc:sqlserver://server:1433;databaseName=databasename;user=username;password=password;autoReconnect=true;"
|
||||
statement => [ "INSERT INTO log (host, timestamp, message) VALUES(?, ?, ?)", "host", "@timestamp", "message" ]
|
||||
}
|
||||
}
|
||||
```
|
||||
17
examples/sqlite.md
Normal file
17
examples/sqlite.md
Normal file
@@ -0,0 +1,17 @@
|
||||
# Example: SQLite3
|
||||
* Tested using https://bitbucket.org/xerial/sqlite-jdbc
|
||||
* SQLite setup - `echo "CREATE table log (host text, timestamp datetime, message text);" | sqlite3 test.db`
|
||||
```
|
||||
input
|
||||
{
|
||||
stdin { }
|
||||
}
|
||||
output {
|
||||
stdout { }
|
||||
|
||||
jdbc {
|
||||
connection_string => 'jdbc:sqlite:test.db'
|
||||
statement => [ "INSERT INTO log (host, timestamp, message) VALUES(?, ?, ?)", "host", "@timestamp", "message" ]
|
||||
}
|
||||
}
|
||||
```
|
||||
5
lib/logstash-output-jdbc_jars.rb
Normal file
5
lib/logstash-output-jdbc_jars.rb
Normal file
@@ -0,0 +1,5 @@
|
||||
# encoding: utf-8
|
||||
require 'logstash/environment'
|
||||
|
||||
root_dir = File.expand_path(File.join(File.dirname(__FILE__), ".."))
|
||||
LogStash::Environment.load_runtime_jars! File.join(root_dir, "vendor")
|
||||
@@ -1,156 +1,322 @@
|
||||
# encoding: utf-8
|
||||
require "logstash/outputs/base"
|
||||
require "logstash/namespace"
|
||||
require "stud/buffer"
|
||||
require "java"
|
||||
require 'logstash/outputs/base'
|
||||
require 'logstash/namespace'
|
||||
require 'concurrent'
|
||||
require 'stud/interval'
|
||||
require 'java'
|
||||
require 'logstash-output-jdbc_jars'
|
||||
|
||||
# Write events to a SQL engine, using JDBC.
|
||||
#
|
||||
# It is upto the user of the plugin to correctly configure the plugin. This
|
||||
# includes correctly crafting the SQL statement, and matching the number of
|
||||
# parameters correctly.
|
||||
class LogStash::Outputs::Jdbc < LogStash::Outputs::Base
|
||||
# Adds buffer support
|
||||
include Stud::Buffer
|
||||
declare_threadsafe! if self.respond_to?(:declare_threadsafe!)
|
||||
|
||||
config_name "jdbc"
|
||||
milestone 1
|
||||
STRFTIME_FMT = '%Y-%m-%d %T.%L'.freeze
|
||||
|
||||
# Driver class
|
||||
config :driver_class, :validate => :string
|
||||
RETRYABLE_SQLSTATE_CLASSES = [
|
||||
# Classes of retryable SQLSTATE codes
|
||||
# Not all in the class will be retryable. However, this is the best that
|
||||
# we've got right now.
|
||||
# If a custom state code is required, set it in retry_sql_states.
|
||||
'08', # Connection Exception
|
||||
'24', # Invalid Cursor State (Maybe retry-able in some circumstances)
|
||||
'25', # Invalid Transaction State
|
||||
'40', # Transaction Rollback
|
||||
'53', # Insufficient Resources
|
||||
'54', # Program Limit Exceeded (MAYBE)
|
||||
'55', # Object Not In Prerequisite State
|
||||
'57', # Operator Intervention
|
||||
'58', # System Error
|
||||
].freeze
|
||||
|
||||
# connection string
|
||||
config :connection_string, :validate => :string, :required => true
|
||||
config_name 'jdbc'
|
||||
|
||||
# Driver class - Reintroduced for https://github.com/theangryangel/logstash-output-jdbc/issues/26
|
||||
config :driver_class, validate: :string
|
||||
|
||||
# Does the JDBC driver support autocommit?
|
||||
config :driver_auto_commit, validate: :boolean, default: true, required: true
|
||||
|
||||
# Where to find the jar
|
||||
# Defaults to not required, and to the original behaviour
|
||||
config :driver_jar_path, validate: :string, required: false
|
||||
|
||||
# jdbc connection string
|
||||
config :connection_string, validate: :string, required: true
|
||||
|
||||
# jdbc username - optional, maybe in the connection string
|
||||
config :username, validate: :string, required: false
|
||||
|
||||
# jdbc password - optional, maybe in the connection string
|
||||
config :password, validate: :string, required: false
|
||||
|
||||
# [ "insert into table (message) values(?)", "%{message}" ]
|
||||
config :statement, :validate => :array, :required => true
|
||||
config :statement, validate: :array, required: true
|
||||
|
||||
# If this is an unsafe statement, use event.sprintf
|
||||
# This also has potential performance penalties due to having to create a
|
||||
# new statement for each event, rather than adding to the batch and issuing
|
||||
# multiple inserts in 1 go
|
||||
config :unsafe_statement, validate: :boolean, default: false
|
||||
|
||||
# Number of connections in the pool to maintain
|
||||
config :max_pool_size, validate: :number, default: 24
|
||||
|
||||
# Connection timeout
|
||||
config :connection_timeout, validate: :number, default: 10000
|
||||
|
||||
# We buffer a certain number of events before flushing that out to SQL.
|
||||
# This setting controls how many events will be buffered before sending a
|
||||
# batch of events.
|
||||
config :flush_size, :validate => :number, :default => 1000
|
||||
config :flush_size, validate: :number, default: 1000
|
||||
|
||||
# The amount of time since last flush before a flush is forced.
|
||||
#
|
||||
# This setting helps ensure slow event rates don't get stuck in Logstash.
|
||||
# For example, if your `flush_size` is 100, and you have received 10 events,
|
||||
# and it has been more than `idle_flush_time` seconds since the last flush,
|
||||
# Logstash will flush those 10 events automatically.
|
||||
#
|
||||
# This helps keep both fast and slow log streams moving along in
|
||||
# a timely manner.
|
||||
#
|
||||
# If you change this value please ensure that you change
|
||||
# max_repeat_exceptions_time accordingly.
|
||||
config :idle_flush_time, :validate => :number, :default => 1
|
||||
# Set initial interval in seconds between retries. Doubled on each retry up to `retry_max_interval`
|
||||
config :retry_initial_interval, validate: :number, default: 2
|
||||
|
||||
# Maximum number of repeating (sequential) exceptions, before we stop retrying
|
||||
# Maximum time between retries, in seconds
|
||||
config :retry_max_interval, validate: :number, default: 128
|
||||
|
||||
# Any additional custom, retryable SQL state codes.
|
||||
# Suitable for configuring retryable custom JDBC SQL state codes.
|
||||
config :retry_sql_states, validate: :array, default: []
|
||||
|
||||
# Maximum number of sequential failed attempts, before we stop retrying.
|
||||
# If set to < 1, then it will infinitely retry.
|
||||
config :max_repeat_exceptions, :validate => :number, :default => 5
|
||||
# At the default values this is a little over 10 minutes
|
||||
config :max_flush_exceptions, validate: :number, default: 10
|
||||
|
||||
# The max number of seconds since the last exception, before we consider it
|
||||
# a different cause.
|
||||
# This value should be carefully considered in respect to idle_flush_time.
|
||||
config :max_repeat_exceptions_time, :validate => :number, :default => 30
|
||||
config :max_repeat_exceptions, obsolete: 'This has been replaced by max_flush_exceptions - which behaves slightly differently. Please check the documentation.'
|
||||
config :max_repeat_exceptions_time, obsolete: 'This is no longer required'
|
||||
config :idle_flush_time, obsolete: 'No longer necessary under Logstash v5'
|
||||
|
||||
public
|
||||
def register
|
||||
@logger.info("JDBC - Starting up")
|
||||
@logger.info('JDBC - Starting up')
|
||||
|
||||
jarpath = File.join(File.dirname(__FILE__), "../../../vendor/jar/jdbc/*.jar")
|
||||
@logger.info(jarpath)
|
||||
Dir[jarpath].each do |jar|
|
||||
@logger.debug("JDBC - Loaded jar", :jar => jar)
|
||||
require jar
|
||||
LogStash::Logger.setup_log4j(@logger)
|
||||
load_jar_files!
|
||||
|
||||
@stopping = Concurrent::AtomicBoolean.new(false)
|
||||
|
||||
@logger.warn('JDBC - Flush size is set to > 1000') if @flush_size > 1000
|
||||
|
||||
if @statement.empty?
|
||||
@logger.error('JDBC - No statement provided. Configuration error.')
|
||||
end
|
||||
|
||||
import @driver_class
|
||||
|
||||
driver = Object.const_get(@driver_class[@driver_class.rindex('.') + 1, @driver_class.length]).new
|
||||
@connection = driver.connect(@connection_string, java.util.Properties.new)
|
||||
|
||||
@logger.debug("JDBC - Created connection", :driver => driver, :connection => @connection)
|
||||
|
||||
if (@flush_size > 1000)
|
||||
@logger.warn("JDBC - Flush size is set to > 1000. May have performance penalties, depending on your SQL engine.")
|
||||
if !@unsafe_statement && @statement.length < 2
|
||||
@logger.error("JDBC - Statement has no parameters. No events will be inserted into SQL as you're not passing any event data. Likely configuration error.")
|
||||
end
|
||||
|
||||
@repeat_exception_count = 0
|
||||
@last_exception_time = Time.now
|
||||
setup_and_test_pool!
|
||||
end
|
||||
|
||||
if (@max_repeat_exceptions > 0) and ((@idle_flush_time * @max_repeat_exceptions) > @max_repeat_exceptions_time)
|
||||
@logger.warn("JDBC - max_repeat_exceptions_time is set such that it may still permit a looping exception. You probably changed idle_flush_time. Considering increasing max_repeat_exceptions_time.")
|
||||
def multi_receive(events)
|
||||
events.each_slice(@flush_size) do |slice|
|
||||
retrying_submit(slice)
|
||||
end
|
||||
|
||||
buffer_initialize(
|
||||
:max_items => @flush_size,
|
||||
:max_interval => @idle_flush_time,
|
||||
:logger => @logger
|
||||
)
|
||||
end
|
||||
|
||||
def receive(event)
|
||||
return unless output?(event)
|
||||
return unless @statement.length > 0
|
||||
|
||||
buffer_receive(event)
|
||||
retrying_submit([event])
|
||||
end
|
||||
|
||||
def flush(events, teardown=false)
|
||||
statement = @connection.prepareStatement(@statement[0])
|
||||
|
||||
events.each do |event|
|
||||
next if @statement.length < 2
|
||||
|
||||
@statement[1..-1].each_with_index do |i, idx|
|
||||
case event[i]
|
||||
when Time
|
||||
# Most reliable solution, cross JDBC driver
|
||||
statement.setString(idx + 1, event[i].iso8601())
|
||||
when Fixnum, Integer
|
||||
statement.setInt(idx + 1, event[i])
|
||||
when Float
|
||||
statement.setFloat(idx + 1, event[i])
|
||||
when String
|
||||
statement.setString(idx + 1, event[i])
|
||||
else
|
||||
statement.setString(idx + 1, event.sprintf(i))
|
||||
end
|
||||
end
|
||||
|
||||
statement.addBatch()
|
||||
end
|
||||
|
||||
begin
|
||||
@logger.debug("JDBC - Sending SQL", :sql => statement.toString())
|
||||
statement.executeBatch()
|
||||
rescue => e
|
||||
# Raising an exception will incur a retry from Stud::Buffer.
|
||||
# Since the exceutebatch failed this should mean any events failed to be
|
||||
# inserted will be re-run. We're going to log it for the lols anyway.
|
||||
@logger.warn("JDBC - Exception. Will automatically retry", :exception => e)
|
||||
end
|
||||
|
||||
statement.close()
|
||||
end
|
||||
|
||||
def on_flush_error(e)
|
||||
return if @max_repeat_exceptions < 1
|
||||
|
||||
if @last_exception == e.to_s
|
||||
@repeat_exception_count += 1
|
||||
else
|
||||
@repeat_exception_count = 0
|
||||
end
|
||||
|
||||
if (@repeat_exception_count >= @max_repeat_exceptions) and (Time.now - @last_exception_time) < @max_repeat_exceptions_time
|
||||
@logger.error("JDBC - Exception repeated more than the maximum configured", :exception => e, :max_repeat_exceptions => @max_repeat_exceptions, :max_repeat_exceptions_time => @max_repeat_exceptions_time)
|
||||
raise e
|
||||
end
|
||||
|
||||
@last_exception_time = Time.now
|
||||
@last_exception = e.to_s
|
||||
end
|
||||
|
||||
def teardown
|
||||
buffer_flush(:final => true)
|
||||
@connection.close()
|
||||
def close
|
||||
@stopping.make_true
|
||||
@pool.close
|
||||
super
|
||||
end
|
||||
|
||||
private
|
||||
|
||||
def setup_and_test_pool!
|
||||
# Setup pool
|
||||
@pool = Java::ComZaxxerHikari::HikariDataSource.new
|
||||
|
||||
@pool.setAutoCommit(@driver_auto_commit)
|
||||
@pool.setDriverClassName(@driver_class) if @driver_class
|
||||
|
||||
@pool.setJdbcUrl(@connection_string)
|
||||
|
||||
@pool.setUsername(@username) if @username
|
||||
@pool.setPassword(@password) if @password
|
||||
|
||||
@pool.setMaximumPoolSize(@max_pool_size)
|
||||
@pool.setConnectionTimeout(@connection_timeout)
|
||||
|
||||
validate_connection_timeout = (@connection_timeout / 1000) / 2
|
||||
|
||||
# Test connection
|
||||
test_connection = @pool.getConnection
|
||||
unless test_connection.isValid(validate_connection_timeout)
|
||||
@logger.error('JDBC - Connection is not valid. Please check connection string or that your JDBC endpoint is available.')
|
||||
end
|
||||
test_connection.close
|
||||
end
|
||||
|
||||
def load_jar_files!
|
||||
# Load jar from driver path
|
||||
unless @driver_jar_path.nil?
|
||||
raise LogStash::ConfigurationError, 'JDBC - Could not find jar file at given path. Check config.' unless File.exist? @driver_jar_path
|
||||
require @driver_jar_path
|
||||
return
|
||||
end
|
||||
|
||||
# Revert original behaviour of loading from vendor directory
|
||||
# if no path given
|
||||
jarpath = if ENV['LOGSTASH_HOME']
|
||||
File.join(ENV['LOGSTASH_HOME'], '/vendor/jar/jdbc/*.jar')
|
||||
else
|
||||
File.join(File.dirname(__FILE__), '../../../vendor/jar/jdbc/*.jar')
|
||||
end
|
||||
|
||||
@logger.debug('JDBC - jarpath', path: jarpath)
|
||||
|
||||
jars = Dir[jarpath]
|
||||
raise LogStash::ConfigurationError, 'JDBC - No jars found. Have you read the README?' if jars.empty?
|
||||
|
||||
jars.each do |jar|
|
||||
@logger.debug('JDBC - Loaded jar', jar: jar)
|
||||
require jar
|
||||
end
|
||||
end
|
||||
|
||||
def submit(events)
|
||||
connection = nil
|
||||
statement = nil
|
||||
events_to_retry = []
|
||||
|
||||
begin
|
||||
connection = @pool.getConnection
|
||||
rescue => e
|
||||
log_jdbc_exception(e, true)
|
||||
# If a connection is not available, then the server has gone away
|
||||
# We're not counting that towards our retry count.
|
||||
return events, false
|
||||
end
|
||||
|
||||
events.each do |event|
|
||||
begin
|
||||
statement = connection.prepareStatement(
|
||||
(@unsafe_statement == true) ? event.sprintf(@statement[0]) : @statement[0]
|
||||
)
|
||||
statement = add_statement_event_params(statement, event) if @statement.length > 1
|
||||
statement.execute
|
||||
rescue => e
|
||||
if retry_exception?(e)
|
||||
events_to_retry.push(event)
|
||||
end
|
||||
ensure
|
||||
statement.close unless statement.nil?
|
||||
end
|
||||
end
|
||||
|
||||
connection.close unless connection.nil?
|
||||
|
||||
return events_to_retry, true
|
||||
end
|
||||
|
||||
def retrying_submit(actions)
|
||||
# Initially we submit the full list of actions
|
||||
submit_actions = actions
|
||||
count_as_attempt = true
|
||||
|
||||
attempts = 1
|
||||
|
||||
sleep_interval = @retry_initial_interval
|
||||
while @stopping.false? and (submit_actions and !submit_actions.empty?)
|
||||
return if !submit_actions || submit_actions.empty? # If everything's a success we move along
|
||||
# We retry whatever didn't succeed
|
||||
submit_actions, count_as_attempt = submit(submit_actions)
|
||||
|
||||
# Everything was a success!
|
||||
break if !submit_actions || submit_actions.empty?
|
||||
|
||||
if @max_flush_exceptions > 0 and count_as_attempt == true
|
||||
attempts += 1
|
||||
|
||||
if attempts > @max_flush_exceptions
|
||||
@logger.error("JDBC - max_flush_exceptions has been reached. #{submit_actions.length} events have been unable to be sent to SQL and are being dropped. See previously logged exceptions for details.")
|
||||
break
|
||||
end
|
||||
end
|
||||
|
||||
# If we're retrying the action sleep for the recommended interval
|
||||
# Double the interval for the next time through to achieve exponential backoff
|
||||
Stud.stoppable_sleep(sleep_interval) { @stopping.true? }
|
||||
sleep_interval = next_sleep_interval(sleep_interval)
|
||||
end
|
||||
end
|
||||
|
||||
def add_statement_event_params(statement, event)
|
||||
@statement[1..-1].each_with_index do |i, idx|
|
||||
if i.is_a? String
|
||||
value = event[i]
|
||||
if value.nil? and i =~ /%\{/
|
||||
value = event.sprintf(i)
|
||||
end
|
||||
else
|
||||
value = i
|
||||
end
|
||||
|
||||
case value
|
||||
when Time
|
||||
# See LogStash::Timestamp, below, for the why behind strftime.
|
||||
statement.setString(idx + 1, value.strftime(STRFTIME_FMT))
|
||||
when LogStash::Timestamp
|
||||
# XXX: Using setString as opposed to setTimestamp, because setTimestamp
|
||||
# doesn't behave correctly in some drivers (Known: sqlite)
|
||||
#
|
||||
# Additionally this does not use `to_iso8601`, since some SQL databases
|
||||
# choke on the 'T' in the string (Known: Derby).
|
||||
#
|
||||
# strftime appears to be the most reliable across drivers.
|
||||
statement.setString(idx + 1, value.time.strftime(STRFTIME_FMT))
|
||||
when Fixnum, Integer
|
||||
statement.setInt(idx + 1, value)
|
||||
when Float
|
||||
statement.setFloat(idx + 1, value)
|
||||
when String
|
||||
statement.setString(idx + 1, value)
|
||||
when true, false
|
||||
statement.setBoolean(idx + 1, value)
|
||||
else
|
||||
statement.setString(idx + 1, nil)
|
||||
end
|
||||
end
|
||||
|
||||
statement
|
||||
end
|
||||
|
||||
def retry_exception?(exception)
|
||||
retrying = (exception.respond_to? 'getSQLState' and (RETRYABLE_SQLSTATE_CLASSES.include?(exception.getSQLState.to_s[0,2]) or @retry_sql_states.include?(exception.getSQLState)))
|
||||
log_jdbc_exception(exception, retrying)
|
||||
|
||||
retrying
|
||||
end
|
||||
|
||||
def log_jdbc_exception(exception, retrying)
|
||||
current_exception = exception
|
||||
log_text = 'JDBC - Exception. ' + (retrying ? 'Retrying' : 'Not retrying') + '.'
|
||||
log_method = (retrying ? 'warn' : 'error')
|
||||
|
||||
loop do
|
||||
@logger.send(log_method, log_text, :exception => current_exception, :backtrace => current_exception.backtrace)
|
||||
|
||||
if current_exception.respond_to? 'getNextException'
|
||||
current_exception = current_exception.getNextException()
|
||||
else
|
||||
current_exception = nil
|
||||
end
|
||||
|
||||
break if current_exception == nil
|
||||
end
|
||||
end
|
||||
|
||||
def next_sleep_interval(current_interval)
|
||||
doubled = current_interval * 2
|
||||
doubled > @retry_max_interval ? @retry_max_interval : doubled
|
||||
end
|
||||
end # class LogStash::Outputs::jdbc
|
||||
|
||||
38
logstash-output-jdbc.gemspec
Normal file
38
logstash-output-jdbc.gemspec
Normal file
@@ -0,0 +1,38 @@
|
||||
Gem::Specification.new do |s|
|
||||
s.name = 'logstash-output-jdbc'
|
||||
s.version = "0.3.0"
|
||||
s.licenses = [ "Apache License (2.0)" ]
|
||||
s.summary = "This plugin allows you to output to SQL, via JDBC"
|
||||
s.description = "This gem is a logstash plugin required to be installed on top of the Logstash core pipeline using $LS_HOME/bin/plugin install gemname. This gem is not a stand-alone program"
|
||||
s.authors = ["the_angry_angel"]
|
||||
s.email = "karl+github@theangryangel.co.uk"
|
||||
s.homepage = "https://github.com/theangryangel/logstash-output-jdbc"
|
||||
s.require_paths = [ "lib" ]
|
||||
|
||||
# Java only
|
||||
s.platform = 'java'
|
||||
|
||||
# Files
|
||||
s.files = Dir.glob('{lib,spec}/**/*.rb') + Dir.glob('vendor/**/*') + %w(LICENSE.txt README.md)
|
||||
|
||||
# Tests
|
||||
s.test_files = s.files.grep(%r{^(test|spec|features)/})
|
||||
|
||||
# Special flag to let us know this is actually a logstash plugin
|
||||
s.metadata = { "logstash_plugin" => "true", "logstash_group" => "output" }
|
||||
|
||||
# Gem dependencies
|
||||
s.add_runtime_dependency 'logstash-core-plugin-api', '~> 1.0'
|
||||
s.add_runtime_dependency 'stud'
|
||||
s.add_runtime_dependency 'logstash-codec-plain'
|
||||
|
||||
s.requirements << "jar 'com.zaxxer:HikariCP', '2.4.2'"
|
||||
s.requirements << "jar 'org.slf4j:slf4j-log4j12', '1.7.21'"
|
||||
|
||||
s.add_development_dependency 'jar-dependencies'
|
||||
s.add_development_dependency 'ruby-maven', '~> 3.3'
|
||||
|
||||
s.add_development_dependency 'logstash-devutils'
|
||||
|
||||
s.add_development_dependency 'rubocop'
|
||||
end
|
||||
8
scripts/travis-before_script.sh
Executable file
8
scripts/travis-before_script.sh
Executable file
@@ -0,0 +1,8 @@
|
||||
#!/bin/bash
|
||||
wget http://search.maven.org/remotecontent?filepath=org/apache/derby/derby/10.12.1.1/derby-10.12.1.1.jar -O /tmp/derby.jar
|
||||
|
||||
sudo apt-get install mysql-server -qq -y
|
||||
echo "create database logstash_output_jdbc_test;" | mysql -u root
|
||||
|
||||
wget http://search.maven.org/remotecontent?filepath=mysql/mysql-connector-java/5.1.38/mysql-connector-java-5.1.38.jar -O /tmp/mysql.jar
|
||||
wget http://search.maven.org/remotecontent?filepath=org/xerial/sqlite-jdbc/3.8.11.2/sqlite-jdbc-3.8.11.2.jar -O /tmp/sqlite.jar
|
||||
3
scripts/travis-variables.sh
Normal file
3
scripts/travis-variables.sh
Normal file
@@ -0,0 +1,3 @@
|
||||
export JDBC_DERBY_JAR=/tmp/derby.jar
|
||||
export JDBC_MYSQL_JAR=/tmp/mysql.jar
|
||||
export JDBC_SQLITE_JAR=/tmp/sqlite.jar
|
||||
135
spec/jdbc_spec_helper.rb
Normal file
135
spec/jdbc_spec_helper.rb
Normal file
@@ -0,0 +1,135 @@
|
||||
require 'logstash/devutils/rspec/spec_helper'
|
||||
require 'logstash/outputs/jdbc'
|
||||
require 'stud/temporary'
|
||||
require 'java'
|
||||
require 'securerandom'
|
||||
|
||||
RSpec.shared_context 'rspec setup' do
|
||||
it 'ensure jar is available' do
|
||||
expect(ENV[jdbc_jar_env]).not_to be_nil, "#{jdbc_jar_env} not defined, required to run tests"
|
||||
expect(File.exist?(ENV[jdbc_jar_env])).to eq(true), "#{jdbc_jar_env} defined, but not valid"
|
||||
end
|
||||
end
|
||||
|
||||
RSpec.shared_context 'when initializing' do
|
||||
it 'shouldn\'t register with a missing jar file' do
|
||||
jdbc_settings['driver_jar_path'] = nil
|
||||
plugin = LogStash::Plugin.lookup('output', 'jdbc').new(jdbc_settings)
|
||||
expect { plugin.register }.to raise_error(LogStash::ConfigurationError)
|
||||
end
|
||||
end
|
||||
|
||||
RSpec.shared_context 'when outputting messages' do
|
||||
let(:logger) { double("logger") }
|
||||
|
||||
let(:jdbc_test_table) do
|
||||
'logstash_output_jdbc_test'
|
||||
end
|
||||
|
||||
let(:jdbc_drop_table) do
|
||||
"DROP TABLE #{jdbc_test_table}"
|
||||
end
|
||||
|
||||
let(:jdbc_create_table) do
|
||||
"CREATE table #{jdbc_test_table} (created_at datetime not null, message varchar(512) not null, message_sprintf varchar(512) not null, static_int int not null, static_bit bit not null)"
|
||||
end
|
||||
|
||||
let(:jdbc_statement) do
|
||||
["insert into #{jdbc_test_table} (created_at, message, message_sprintf, static_int, static_bit) values(?, ?, ?, ?, ?)", '@timestamp', 'message', 'sprintf-%{message}', 1, true]
|
||||
end
|
||||
|
||||
let(:systemd_database_service) do
|
||||
nil
|
||||
end
|
||||
|
||||
let(:event_fields) do
|
||||
{ 'message' => "test-message #{SecureRandom.uuid}" }
|
||||
end
|
||||
|
||||
let(:event) { LogStash::Event.new(event_fields) }
|
||||
|
||||
let(:plugin) do
|
||||
# Setup plugin
|
||||
output = LogStash::Plugin.lookup('output', 'jdbc').new(jdbc_settings)
|
||||
output.register
|
||||
output.logger = logger
|
||||
|
||||
# Setup table
|
||||
c = output.instance_variable_get(:@pool).getConnection
|
||||
|
||||
# Derby doesn't support IF EXISTS.
|
||||
# Seems like the quickest solution. Bleurgh.
|
||||
begin
|
||||
stmt = c.createStatement
|
||||
stmt.executeUpdate(jdbc_drop_table)
|
||||
rescue
|
||||
# noop
|
||||
ensure
|
||||
stmt.close
|
||||
|
||||
stmt = c.createStatement
|
||||
stmt.executeUpdate(jdbc_create_table)
|
||||
stmt.close
|
||||
c.close
|
||||
end
|
||||
|
||||
output
|
||||
end
|
||||
|
||||
it 'should save a event' do
|
||||
expect { plugin.multi_receive([event]) }.to_not raise_error
|
||||
|
||||
# Verify the number of items in the output table
|
||||
c = plugin.instance_variable_get(:@pool).getConnection
|
||||
stmt = c.prepareStatement("select count(*) as total from #{jdbc_test_table} where message = ?")
|
||||
stmt.setString(1, event['message'])
|
||||
rs = stmt.executeQuery
|
||||
count = 0
|
||||
count = rs.getInt('total') while rs.next
|
||||
stmt.close
|
||||
c.close
|
||||
|
||||
expect(count).to eq(1)
|
||||
end
|
||||
|
||||
it 'should not save event, and log an unretryable exception' do
|
||||
e = LogStash::Event.new({})
|
||||
|
||||
expect(logger).to receive(:error).once.with(/JDBC - Exception. Not retrying/, Hash)
|
||||
expect { plugin.multi_receive([e]) }.to_not raise_error
|
||||
end
|
||||
|
||||
it 'it should retry after a connection loss, and log a warning' do
|
||||
skip "does not run as a service" if systemd_database_service.nil?
|
||||
|
||||
p = plugin
|
||||
|
||||
# Check that everything is fine right now
|
||||
expect { p.multi_receive([event]) }.not_to raise_error
|
||||
|
||||
# Start a thread to stop and restart the service.
|
||||
t = Thread.new(systemd_database_service) { |systemd_database_service|
|
||||
start_stop_cmd = 'sudo /etc/init.d/%<service>s* %<action>s'
|
||||
|
||||
`which systemctl`
|
||||
if $?.success?
|
||||
start_stop_cmd = 'sudo systemctl %<action>s %<service>s'
|
||||
end
|
||||
|
||||
cmd = start_stop_cmd % { action: 'stop', service: systemd_database_service }
|
||||
`#{cmd}`
|
||||
sleep 10
|
||||
|
||||
cmd = start_stop_cmd % { action: 'start', service: systemd_database_service }
|
||||
`#{cmd}`
|
||||
}
|
||||
|
||||
# Wait a few seconds to the service to stop
|
||||
sleep 5
|
||||
|
||||
expect(logger).to receive(:warn).at_least(:once).with(/JDBC - Exception. Retrying/, Hash)
|
||||
expect { p.multi_receive([event]) }.to_not raise_error
|
||||
|
||||
t.join
|
||||
end
|
||||
end
|
||||
25
spec/outputs/jdbc_derby_spec.rb
Normal file
25
spec/outputs/jdbc_derby_spec.rb
Normal file
@@ -0,0 +1,25 @@
|
||||
require_relative '../jdbc_spec_helper'
|
||||
|
||||
describe 'logstash-output-jdbc: derby', if: ENV['JDBC_DERBY_JAR'] do
|
||||
include_context 'rspec setup'
|
||||
include_context 'when initializing'
|
||||
include_context 'when outputting messages'
|
||||
|
||||
let(:jdbc_jar_env) do
|
||||
'JDBC_DERBY_JAR'
|
||||
end
|
||||
|
||||
let(:jdbc_create_table) do
|
||||
"CREATE table #{jdbc_test_table} (created_at timestamp not null, message varchar(512) not null, message_sprintf varchar(512) not null, static_int int not null, static_bit boolean not null)"
|
||||
end
|
||||
|
||||
let(:jdbc_settings) do
|
||||
{
|
||||
'driver_class' => 'org.apache.derby.jdbc.EmbeddedDriver',
|
||||
'connection_string' => 'jdbc:derby:memory:testdb;create=true',
|
||||
'driver_jar_path' => ENV[jdbc_jar_env],
|
||||
'statement' => jdbc_statement,
|
||||
'max_flush_exceptions' => 1
|
||||
}
|
||||
end
|
||||
end
|
||||
25
spec/outputs/jdbc_mysql_spec.rb
Normal file
25
spec/outputs/jdbc_mysql_spec.rb
Normal file
@@ -0,0 +1,25 @@
|
||||
require_relative '../jdbc_spec_helper'
|
||||
|
||||
describe 'logstash-output-jdbc: mysql', if: ENV['JDBC_MYSQL_JAR'] do
|
||||
include_context 'rspec setup'
|
||||
include_context 'when initializing'
|
||||
include_context 'when outputting messages'
|
||||
|
||||
let(:jdbc_jar_env) do
|
||||
'JDBC_MYSQL_JAR'
|
||||
end
|
||||
|
||||
let(:systemd_database_service) do
|
||||
'mysql'
|
||||
end
|
||||
|
||||
let(:jdbc_settings) do
|
||||
{
|
||||
'driver_class' => 'com.mysql.jdbc.Driver',
|
||||
'connection_string' => 'jdbc:mysql://localhost/logstash_output_jdbc_test?user=root',
|
||||
'driver_jar_path' => ENV[jdbc_jar_env],
|
||||
'statement' => jdbc_statement,
|
||||
'max_flush_exceptions' => 1
|
||||
}
|
||||
end
|
||||
end
|
||||
11
spec/outputs/jdbc_spec.rb
Normal file
11
spec/outputs/jdbc_spec.rb
Normal file
@@ -0,0 +1,11 @@
|
||||
require_relative '../jdbc_spec_helper'
|
||||
|
||||
describe LogStash::Outputs::Jdbc do
|
||||
context 'when initializing' do
|
||||
it 'shouldn\'t register without a config' do
|
||||
expect do
|
||||
LogStash::Plugin.lookup('output', 'jdbc').new
|
||||
end.to raise_error(LogStash::ConfigurationError)
|
||||
end
|
||||
end
|
||||
end
|
||||
27
spec/outputs/jdbc_sqlite_spec.rb
Normal file
27
spec/outputs/jdbc_sqlite_spec.rb
Normal file
@@ -0,0 +1,27 @@
|
||||
require_relative '../jdbc_spec_helper'
|
||||
|
||||
describe 'logstash-output-jdbc: sqlite', if: ENV['JDBC_SQLITE_JAR'] do
|
||||
JDBC_SQLITE_FILE = '/tmp/logstash_output_jdbc_test.db'.freeze
|
||||
|
||||
before(:context) do
|
||||
File.delete(JDBC_SQLITE_FILE) if File.exist? JDBC_SQLITE_FILE
|
||||
end
|
||||
|
||||
include_context 'rspec setup'
|
||||
include_context 'when initializing'
|
||||
include_context 'when outputting messages'
|
||||
|
||||
let(:jdbc_jar_env) do
|
||||
'JDBC_SQLITE_JAR'
|
||||
end
|
||||
|
||||
let(:jdbc_settings) do
|
||||
{
|
||||
'driver_class' => 'org.sqlite.JDBC',
|
||||
'connection_string' => "jdbc:sqlite:#{JDBC_SQLITE_FILE}",
|
||||
'driver_jar_path' => ENV[jdbc_jar_env],
|
||||
'statement' => jdbc_statement,
|
||||
'max_flush_exceptions' => 1
|
||||
}
|
||||
end
|
||||
end
|
||||
Reference in New Issue
Block a user