2014-04-15 11:32:41 +00:00
# encoding: utf-8
require " logstash/outputs/base "
require " logstash/namespace "
2016-05-14 21:17:34 +00:00
require " concurrent "
require " stud/interval "
2014-08-09 12:56:11 +00:00
require " java "
2015-11-14 20:04:16 +00:00
require " logstash-output-jdbc_jars "
2014-04-15 11:32:41 +00:00
class LogStash :: Outputs :: Jdbc < LogStash :: Outputs :: Base
2016-05-03 14:28:01 +00:00
STRFTIME_FMT = " %Y-%m-%d %T.%L " . freeze
2014-04-15 11:32:41 +00:00
config_name " jdbc "
2015-12-23 09:42:53 +00:00
# Driver class - Reintroduced for https://github.com/theangryangel/logstash-output-jdbc/issues/26
config :driver_class , :validate = > :string
# Does the JDBC driver support autocommit?
config :driver_auto_commit , :validate = > :boolean , :default = > true , :required = > true
2014-04-15 11:32:41 +00:00
2015-11-14 20:04:16 +00:00
# Where to find the jar
# Defaults to not required, and to the original behaviour
config :driver_jar_path , :validate = > :string , :required = > false
# jdbc connection string
2014-04-15 11:32:41 +00:00
config :connection_string , :validate = > :string , :required = > true
2015-11-14 20:04:16 +00:00
# jdbc username - optional, maybe in the connection string
config :username , :validate = > :string , :required = > false
# jdbc password - optional, maybe in the connection string
config :password , :validate = > :string , :required = > false
2014-08-09 12:56:11 +00:00
# [ "insert into table (message) values(?)", "%{message}" ]
2014-04-15 11:32:41 +00:00
config :statement , :validate = > :array , :required = > true
2014-05-31 13:33:59 +00:00
2015-11-15 12:35:57 +00:00
# If this is an unsafe statement, use event.sprintf
# This also has potential performance penalties due to having to create a
# new statement for each event, rather than adding to the batch and issuing
# multiple inserts in 1 go
config :unsafe_statement , :validate = > :boolean , :default = > false
2015-11-14 20:04:16 +00:00
# Number of connections in the pool to maintain
config :max_pool_size , :validate = > :number , :default = > 5
# Connection timeout
2016-02-16 15:29:08 +00:00
config :connection_timeout , :validate = > :number , :default = > 10000
2015-11-14 20:04:16 +00:00
2014-06-01 12:16:33 +00:00
# We buffer a certain number of events before flushing that out to SQL.
# This setting controls how many events will be buffered before sending a
# batch of events.
2014-05-31 13:33:59 +00:00
config :flush_size , :validate = > :number , :default = > 1000
2016-05-14 21:17:34 +00:00
# Set initial interval in seconds between retries. Doubled on each retry up to `retry_max_interval`
config :retry_initial_interval , :validate = > :number , :default = > 2
# Maximum time between retries, in seconds
config :retry_max_interval , :validate = > :number , :default = > 128
# Maximum number of sequential failed attempts, before we stop retrying.
2014-08-09 12:56:11 +00:00
# If set to < 1, then it will infinitely retry.
2015-11-18 10:06:11 +00:00
config :max_flush_exceptions , :validate = > :number , :default = > 0
2014-08-09 12:56:11 +00:00
2015-11-18 10:06:11 +00:00
config :max_repeat_exceptions , :obsolete = > " This has been replaced by max_flush_exceptions - which behaves slightly differently. Please check the documentation. "
config :max_repeat_exceptions_time , :obsolete = > " This is no longer required "
2016-05-14 21:17:34 +00:00
config :idle_flush_time , :obsolete = > " No longer necessary under Logstash v5 "
2014-08-09 12:56:11 +00:00
2014-04-15 11:32:41 +00:00
public
def register
2014-08-09 12:56:11 +00:00
@logger . info ( " JDBC - Starting up " )
2014-04-15 11:32:41 +00:00
2016-05-13 21:12:57 +00:00
LogStash :: Logger . setup_log4j ( @logger )
2015-11-14 20:04:16 +00:00
load_jar_files!
2014-04-15 11:32:41 +00:00
2016-05-14 21:17:34 +00:00
@stopping = Concurrent :: AtomicBoolean . new ( false )
2015-11-18 10:06:11 +00:00
2014-05-31 13:33:59 +00:00
if ( @flush_size > 1000 )
2015-11-14 20:04:16 +00:00
@logger . warn ( " JDBC - Flush size is set to > 1000 " )
2014-08-09 12:56:11 +00:00
end
2015-12-23 10:06:50 +00:00
if @statement . length < 1
@logger . error ( " JDBC - No statement provided. Configuration error. " )
end
if ( ! @unsafe_statement and @statement . length < 2 )
@logger . error ( " JDBC - Statement has no parameters. No events will be inserted into SQL as you're not passing any event data. Likely configuration error. " )
end
2015-12-30 12:04:23 +00:00
setup_and_test_pool!
2014-05-31 13:33:59 +00:00
end
2014-04-16 15:41:45 +00:00
2016-05-14 21:17:34 +00:00
def multi_receive ( events )
events . each_slice ( @flush_size ) do | slice |
retrying_submit ( slice )
2014-08-09 12:56:11 +00:00
end
end
2014-04-16 15:41:45 +00:00
def teardown
2015-11-14 20:04:16 +00:00
@pool . close ( )
2014-04-16 15:41:45 +00:00
super
2014-04-15 11:32:41 +00:00
end
2016-05-14 21:17:34 +00:00
def close
@stopping . make_true
end
2015-11-14 20:04:16 +00:00
private
2015-12-30 12:04:23 +00:00
def setup_and_test_pool!
# Setup pool
@pool = Java :: ComZaxxerHikari :: HikariDataSource . new
@pool . setAutoCommit ( @driver_auto_commit )
@pool . setDriverClassName ( @driver_class ) if @driver_class
@pool . setJdbcUrl ( @connection_string )
@pool . setUsername ( @username ) if @username
@pool . setPassword ( @password ) if @password
@pool . setMaximumPoolSize ( @max_pool_size )
@pool . setConnectionTimeout ( @connection_timeout )
2016-02-16 15:29:08 +00:00
validate_connection_timeout = ( @connection_timeout / 1000 ) / 2
2015-12-30 12:04:23 +00:00
# Test connection
test_connection = @pool . getConnection ( )
2016-02-16 15:29:08 +00:00
unless test_connection . isValid ( validate_connection_timeout )
2015-12-30 12:04:23 +00:00
@logger . error ( " JDBC - Connection is not valid. Please check connection string or that your JDBC endpoint is available. " )
end
test_connection . close ( )
end
2015-11-14 20:04:16 +00:00
def load_jar_files!
# Load jar from driver path
unless @driver_jar_path . nil?
raise Exception . new ( " JDBC - Could not find jar file at given path. Check config. " ) unless File . exists? @driver_jar_path
require @driver_jar_path
return
end
# Revert original behaviour of loading from vendor directory
# if no path given
if ENV [ 'LOGSTASH_HOME' ]
jarpath = File . join ( ENV [ 'LOGSTASH_HOME' ] , " /vendor/jar/jdbc/*.jar " )
else
jarpath = File . join ( File . dirname ( __FILE__ ) , " ../../../vendor/jar/jdbc/*.jar " )
end
@logger . debug ( " JDBC - jarpath " , path : jarpath )
jars = Dir [ jarpath ]
raise Exception . new ( " JDBC - No jars found in jarpath. Have you read the README? " ) if jars . empty?
jars . each do | jar |
@logger . debug ( " JDBC - Loaded jar " , :jar = > jar )
require jar
end
end
2015-11-15 12:35:57 +00:00
2016-05-14 21:17:34 +00:00
def submit_safe ( events )
2016-04-16 13:33:30 +00:00
connection = nil
statement = nil
2016-05-14 21:17:34 +00:00
events_to_retry = [ ]
2016-04-16 13:33:30 +00:00
begin
connection = @pool . getConnection ( )
2016-05-14 21:17:34 +00:00
rescue = > e
log_jdbc_exception ( e )
return events
end
2015-11-15 12:35:57 +00:00
2016-05-14 21:17:34 +00:00
begin
statement = connection . prepareStatement ( @statement [ 0 ] )
2016-04-16 13:33:30 +00:00
events . each do | event |
next if event . cancelled?
next if @statement . length < 2
statement = add_statement_event_params ( statement , event )
2015-11-15 12:35:57 +00:00
2016-04-16 13:33:30 +00:00
statement . addBatch ( )
end
2015-11-15 12:35:57 +00:00
statement . executeBatch ( )
2016-05-14 21:17:34 +00:00
rescue java . sql . BatchUpdateException = > e
# Only retry the failed items from the batch
updates = e . getUpdateCounts ( )
events_to_retry = events
updates . each_with_index { | update , idx |
if ( update != java . sql . Statement . EXECUTE_FAILED )
# Remove any successful events
events_to_retry [ idx ] = nil
end
}
# Remove the nil entries
events_to_retry = events_to_retry . compact
log_jdbc_exception ( e )
2015-11-15 12:35:57 +00:00
rescue = > e
2016-05-14 21:17:34 +00:00
events_to_retry = events
2015-11-17 10:32:16 +00:00
log_jdbc_exception ( e )
2015-11-15 12:35:57 +00:00
ensure
2016-04-16 13:33:30 +00:00
statement . close ( ) unless statement . nil?
2015-11-15 12:35:57 +00:00
end
2016-05-14 21:17:34 +00:00
connection . close ( ) unless connection . nil?
return events_to_retry
2015-11-15 12:35:57 +00:00
end
2016-05-14 21:17:34 +00:00
def submit_unsafe ( events )
2016-04-16 13:33:30 +00:00
connection = nil
statement = nil
2016-05-14 21:17:34 +00:00
events_to_retry = [ ]
2016-04-16 13:33:30 +00:00
begin
connection = @pool . getConnection ( )
2016-05-14 21:17:34 +00:00
rescue = > e
log_jdbc_exception ( e )
return events
end
2015-11-15 12:35:57 +00:00
2016-05-14 21:17:34 +00:00
events . each do | event |
begin
2016-04-16 13:33:30 +00:00
statement = connection . prepareStatement ( event . sprintf ( @statement [ 0 ] ) )
statement = add_statement_event_params ( statement , event ) if @statement . length > 1
2015-11-15 12:35:57 +00:00
2015-11-17 10:32:16 +00:00
statement . execute ( )
2016-05-14 21:17:34 +00:00
rescue = > e
log_jdbc_exception ( e )
events_to_retry . push ( event )
ensure
statement . close ( ) unless statement . nil?
statement = nil
end
end
connection . close ( ) unless connection . nil?
return events_to_retry
end
def retrying_submit ( actions )
# Initially we submit the full list of actions
submit_actions = actions
attempts = 0
sleep_interval = @retry_initial_interval
while submit_actions && submit_actions . length > 0
return if ! submit_actions || submit_actions . empty? # If everything's a success we move along
# We retry with whatever is didn't succeed
begin
if @unsafe_statement == true
submit_actions = submit_unsafe ( submit_actions )
else
submit_actions = submit_safe ( submit_actions )
end
rescue = > e
log_jdbc_exception ( e )
end
if @max_flush_exceptions > 0
attempts += 1
2016-04-16 13:33:30 +00:00
2016-05-14 21:17:34 +00:00
if attempts > @max_flush_exceptions
@logger . error ( " JDBC - max_flush_exceptions has been reached " )
break
end
2015-11-17 10:32:16 +00:00
end
2016-05-14 21:17:34 +00:00
# Everything was a success!
break if ! submit_actions || submit_actions . empty?
# If we're retrying the action sleep for the recommended interval
# Double the interval for the next time through to achieve exponential backoff
Stud . stoppable_sleep ( sleep_interval ) { @stopping . true? }
sleep_interval = next_sleep_interval ( sleep_interval )
2015-11-15 12:35:57 +00:00
end
end
def add_statement_event_params ( statement , event )
@statement [ 1 .. - 1 ] . each_with_index do | i , idx |
2016-05-13 21:12:57 +00:00
case event . get ( i )
Fix toString method of LogStash::Timestamp
According to LogStash::Timestamp (https://github.com/elastic/logstash/blob/bb30cc773bcf34d63d634dbfcf2803278a5140b3/logstash-core-event/lib/logstash/timestamp.rb) doesn't support iso8601, which results in error if the timestamp of logstash is used directly.
If should support to_s of to_iso8601.
:message=>"Failed to flush outgoing items", :outgoing_count=>1, :exception=>"NoMethodError", :backtrace=>["/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-output-jdbc-0.2.3/lib/logstash/outputs/jdbc.rb:255:in `add_statement_event_params'", "org/jruby/RubyArray.java:1613:in `each'", "org/jruby/RubyEnumerable.java:974:in `each_with_index'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-output-jdbc-0.2.3/lib/logstash/outputs/jdbc.rb:251:in `add_statement_event_params'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-output-jdbc-0.2.3/lib/logstash/outputs/jdbc.rb:203:in `safe_flush'", "org/jruby/RubyArray.java:1613:in `each'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-output-jdbc-0.2.3/lib/logstash/outputs/jdbc.rb:200:in `safe_flush'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-output-jdbc-0.2.3/lib/logstash/outputs/jdbc.rb:120:in `flush'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/stud-0.0.22/lib/stud/buffer.rb:219:in `buffer_flush'", "org/jruby/RubyHash.java:1342:in `each'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/stud-0.0.22/lib/stud/buffer.rb:216:in `buffer_flush'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/stud-0.0.22/lib/stud/buffer.rb:159:in `buffer_receive'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-output-jdbc-0.2.3/lib/logstash/outputs/jdbc.rb:113:in `receive'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-core-2.3.0-java/lib/logstash/outputs/base.rb:83:in `multi_receive'", "org/jruby/RubyArray.java:1613:in `each'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-core-2.3.0-java/lib/logstash/outputs/base.rb:83:in `multi_receive'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-core-2.3.0-java/lib/logstash/output_delegator.rb:130:in `worker_multi_receive'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-core-2.3.0-java/lib/logstash/output_delegator.rb:114:in `multi_receive'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-core-2.3.0-java/lib/logstash/pipeline.rb:305:in `output_batch'", "org/jruby/RubyHash.java:1342:in `each'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-core-2.3.0-java/lib/logstash/pipeline.rb:305:in `output_batch'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-core-2.3.0-java/lib/logstash/pipeline.rb:236:in `worker_loop'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-core-2.3.0-java/lib/logstash/pipeline.rb:205:in `start_workers'"], :level=>:warn}
2016-04-11 13:19:48 +00:00
when Time
2016-05-03 14:28:01 +00:00
# See LogStash::Timestamp, below, for the why behind strftime.
2016-05-13 21:12:57 +00:00
statement . setString ( idx + 1 , event . get ( i ) . strftime ( STRFTIME_FMT ) )
Fix toString method of LogStash::Timestamp
According to LogStash::Timestamp (https://github.com/elastic/logstash/blob/bb30cc773bcf34d63d634dbfcf2803278a5140b3/logstash-core-event/lib/logstash/timestamp.rb) doesn't support iso8601, which results in error if the timestamp of logstash is used directly.
If should support to_s of to_iso8601.
:message=>"Failed to flush outgoing items", :outgoing_count=>1, :exception=>"NoMethodError", :backtrace=>["/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-output-jdbc-0.2.3/lib/logstash/outputs/jdbc.rb:255:in `add_statement_event_params'", "org/jruby/RubyArray.java:1613:in `each'", "org/jruby/RubyEnumerable.java:974:in `each_with_index'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-output-jdbc-0.2.3/lib/logstash/outputs/jdbc.rb:251:in `add_statement_event_params'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-output-jdbc-0.2.3/lib/logstash/outputs/jdbc.rb:203:in `safe_flush'", "org/jruby/RubyArray.java:1613:in `each'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-output-jdbc-0.2.3/lib/logstash/outputs/jdbc.rb:200:in `safe_flush'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-output-jdbc-0.2.3/lib/logstash/outputs/jdbc.rb:120:in `flush'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/stud-0.0.22/lib/stud/buffer.rb:219:in `buffer_flush'", "org/jruby/RubyHash.java:1342:in `each'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/stud-0.0.22/lib/stud/buffer.rb:216:in `buffer_flush'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/stud-0.0.22/lib/stud/buffer.rb:159:in `buffer_receive'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-output-jdbc-0.2.3/lib/logstash/outputs/jdbc.rb:113:in `receive'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-core-2.3.0-java/lib/logstash/outputs/base.rb:83:in `multi_receive'", "org/jruby/RubyArray.java:1613:in `each'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-core-2.3.0-java/lib/logstash/outputs/base.rb:83:in `multi_receive'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-core-2.3.0-java/lib/logstash/output_delegator.rb:130:in `worker_multi_receive'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-core-2.3.0-java/lib/logstash/output_delegator.rb:114:in `multi_receive'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-core-2.3.0-java/lib/logstash/pipeline.rb:305:in `output_batch'", "org/jruby/RubyHash.java:1342:in `each'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-core-2.3.0-java/lib/logstash/pipeline.rb:305:in `output_batch'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-core-2.3.0-java/lib/logstash/pipeline.rb:236:in `worker_loop'", "/opt/logstash/vendor/bundle/jruby/1.9/gems/logstash-core-2.3.0-java/lib/logstash/pipeline.rb:205:in `start_workers'"], :level=>:warn}
2016-04-11 13:19:48 +00:00
when LogStash :: Timestamp
2016-05-03 14:28:01 +00:00
# XXX: Using setString as opposed to setTimestamp, because setTimestamp
# doesn't behave correctly in some drivers (Known: sqlite)
#
# Additionally this does not use `to_iso8601`, since some SQL databases
# choke on the 'T' in the string (Known: Derby).
#
# strftime appears to be the most reliable across drivers.
2016-05-13 21:12:57 +00:00
statement . setString ( idx + 1 , event . get ( i ) . time . strftime ( STRFTIME_FMT ) )
2015-11-15 12:35:57 +00:00
when Fixnum , Integer
2016-05-13 21:12:57 +00:00
statement . setInt ( idx + 1 , event . get ( i ) )
2015-11-15 12:35:57 +00:00
when Float
2016-05-13 21:12:57 +00:00
statement . setFloat ( idx + 1 , event . get ( i ) )
2015-11-15 12:35:57 +00:00
when String
2016-05-13 21:12:57 +00:00
statement . setString ( idx + 1 , event . get ( i ) )
2015-11-15 12:35:57 +00:00
when true
statement . setBoolean ( idx + 1 , true )
when false
statement . setBoolean ( idx + 1 , false )
else
2016-05-13 21:12:57 +00:00
if event . get ( i ) . nil? and i =~ / % \ { /
2015-11-18 10:06:11 +00:00
statement . setString ( idx + 1 , event . sprintf ( i ) )
else
statement . setString ( idx + 1 , nil )
end
2015-11-15 12:35:57 +00:00
end
end
statement
end
2015-11-17 10:32:16 +00:00
2016-05-02 17:11:27 +00:00
def log_jdbc_exception ( exception )
current_exception = exception
2015-11-17 10:32:16 +00:00
loop do
2016-05-02 17:11:27 +00:00
@logger . error ( " JDBC Exception encountered: Will automatically retry. " , :exception = > current_exception )
current_exception = current_exception . getNextException ( )
break if current_exception == nil
2015-11-17 10:32:16 +00:00
end
end
2016-05-14 21:17:34 +00:00
def next_sleep_interval ( current_interval )
doubled = current_interval * 2
doubled > @retry_max_interval ? @retry_max_interval : doubled
end
2014-04-15 11:32:41 +00:00
end # class LogStash::Outputs::jdbc