All notable changes to this project will be documented in this file.
- Fields
cache_control
,content_disposition
,content_language
andwebsite_redirect_location
added to theaws_s3
output. - Field
cors.enabled
andcors.allowed_origins
added to the server widehttp
config. - For Kafka components the config now supports the
rack_id
field which may contain a rack identifier for the Kafka client. - Allow mapping imports in Bloblang environments to be disabled.
- Go API: Isolated Bloblang environments are now honored by all components.
- Go API: The stream builder now evaluates environment variable interpolations.
- Field
unsafe_dynamic_query
added to thesql
processor. - The
kafka
output now supportszstd
compression.
- The
test
subcommand now expands resource glob patterns (benthos -r "./foo/*.yaml" test ./...
). - The Bloblang equality operator now returns
false
when comparing non-null values withnull
rather than a mismatched types error.
- New experimental
gcp_bigquery
output. - Go API: It's now possible to parse a config spec directly with
ParseYAML
. - Bloblang methods and functions now support named parameters.
- Field
args_mapping
added to thecassandra
output. - For NATS, NATS Streaming and Jetstream components the config now supports specifying either
nkey_file
oruser_credentials_file
to configure authentication.
- The
mqtt
input and output now support sending a last will, configuring a keep alive timeout, and setting retained out output messages. - Go API: New stream builder
AddBatchProducerFunc
andAddBatchConsumerFunc
methods. - Field
gzip_compression
added to theelasticsearch
output. - The
redis_streams
input now supports creating the stream with theMKSTREAM
command (enabled by default). - The
kafka
output now supports manual partition allocation using interpolation functions in the fieldpartition
.
- The bloblang method
contains
now correctly compares numerical values in arrays and objects.
- Go API: Added ability to create and register
BatchBuffer
plugins. - New
system_window
buffer for processing message windows (sliding or tumbling) following the system clock. - Field
root_cas
added to all TLS configuration blocks. - The
sftp
input and output now support key based authentication. - New Bloblang function
nanoid
. - The
gcp_cloud_storage
output now supports custom collision behaviour with the fieldcollision_mode
. - Field
priority
added to theamqp_0_9
output. - Operator
keys
added to theredis
processor. - The
http_client
input when configured in stream mode now allows message body interpolation functions within the URL and header parameters.
- Fixed a panic that would occur when executing a pipeline where processor or input resources reference rate limits.
- The
elasticsearch
output now supports delete, update and index operations. - Go API: Added ability to create and register
BatchInput
plugins.
- Prevented the
http_server
input from blocking graceful pipeline termination indefinitely. - Removed annoying nil error log from HTTP client components when parsing responses.
- The
redis_streams
,redis_pubsub
andredis_list
outputs now all support batching for higher throughput. - The
amqp_1
input and output now support passing and receiving metadata as annotations. - Config unit test definitions can now use files for both the input and expected output.
- Field
track_properties
added to theazure_queue_storage
input for enriching messages with properties such as the message backlog. - Go API: The new plugin APIs, available at
./public/service
, are considered stable. - The streams mode API now uses the setting
http.read_timeout
for timing out stream CRUD endpoints.
- The Bloblang function
random_int
now only resolves dynamic arguments once during the lifetime of the mapping. Documentation has been updated in order to clarify the behaviour with dynamic arguments. - Fixed an issue where plugins registered would return
failed to obtain docs for X type Y
linting errors. - HTTP client components are now more permissive regarding invalid Content-Type headers.
- New CLI flag
--set
(-s
) for overriding arbitrary fields in a config. E.g.-s input.type=http_server
would override the config setting the input type tohttp_server
. - Unit test definitions now support mocking components.
- The
nats
input now supports acks. - The
memory
andfile
cache types now expose metrics akin to other caches.
- The
switch
output whenretry_until_success
is set tofalse
will now provide granular nacks to pre-batched messages. - The URL printed in error messages when HTTP client components fail should now show interpolated values as they were interpreted.
- Go Plugins API V2: Batched processors should now show in tracing, and no longer complain about spans being closed more than once.
- Algorithm
lz4
added to thecompress
anddecompress
processors. - New experimental
aws_dynamodb_partiql
processor. - Go Plugins API: new run opt
OptUseContext
for an extra shutdown mechanism.
- Fixed an issue here the
http_client
would prematurely drop connections when configured withstream.enabled
set totrue
. - Prevented closed output brokers from leaving child outputs running when they've failed to establish a connection.
- Fixed metrics prefixes in streams mode for nested components.
- CLI flag
max-token-length
added to theblobl
subcommand. - Go Plugins API: Plugin components can now be configured seamlessly like native components, meaning the namespace
plugin
is no longer required and configuration fields can be placed within the namespace of the plugin itself. Note that the old style (withinplugin
) is still supported. - The
http_client
input fieldsurl
andheaders
now support interpolation functions that access metadata and contents of the last received message. - Rate limit resources now emit
checked
,limited
anderror
metrics. - A new experimental plugins API is available for early adopters, and can be found at
./public/x/service
. - A new experimental template system is available for early adopters, examples can be found in
./template
. - New beta Bloblang method
bloblang
for executing dynamic mappings. - All
http
components now support a betajwt
authentication mechanism. - New experimental
schema_registry_decode
processor. - New Bloblang method
parse_duration
for parsing duration strings into an integer. - New experimental
twitter_search
input. - New field
args_mapping
added to thesql
processor and output for mapping explicitly typed arguments. - Added format
csv
to theunarchive
processor. - The
redis
processor now supportsincrby
operations. - New experimental
discord
input and output. - The
http_server
input now adds a metadata fieldhttp_server_verb
. - New Bloblang methods
parse_yaml
andformat_yaml
. - CLI flag
env-file
added to Benthos for parsing dotenv files. - New
mssql
SQL driver for thesql
processor and output. - New POST endpoint
/resources/{type}/{id}
added to Benthos streams mode for dynamically mutating resource configs.
- Go Plugins API: The Bloblang
ArgSpec
now returns a public error typeArgError
. - Components that support glob paths (
file
,csv
, etc) now also support super globs (double asterisk). - The
aws_kinesis
input is now stable. - The
gcp_cloud_storage
input and output are now beta. - The
kinesis
input is now deprecated. - Go Plugins API: the minimum version of Go required is now 1.16.
- Fixed a rare panic caused when executing a
workflow
resource processor that referencesbranch
resources across parallel threads. - The
mqtt
input with multiple topics now works with brokers that would previously error on multiple subscriptions. - Fixed initialisation of components configured as resources that reference other resources, where under certain circumstances the components would fail to obtain a true reference to the target resource. This fix makes it so that resources are accessed only when used, which will also make it possible to introduce dynamic resources in future.
- The streams mode endpoint
/streams/{id}/stats
should now work again provided the default manager is used.
- The
branch
processor now writes error logs when the request or result map fails. - The
branch
processor (andworkflow
by proxy) now allow errors to be mapped into the branch usingerror()
in therequest_map
. - Added a linting rule that warns against having a
reject
output under aswitch
broker withoutretry_until_success
disabled. - Prevented a panic or variable corruption that could occur when a Bloblang mapping is executed by parallel threads.
- The
create
subcommand now supports a--small
/-s
flag that reduces the output down to only core components and common fields. - Go Plugins API: Added method
Overlay
to the public Bloblang package. - The
http_server
input now adds path parameters (/{foo}/{bar}
) to the metadata of ingested messages. - The
stdout
output now has acodec
field. - New Bloblang methods
format_timestamp_strftime
andparse_timestamp_strptime
. - New experimental
nats_jetstream
input and output.
- Go Plugins API: Bloblang method and function plugins now automatically resolve dynamic arguments.
- Fixed a regression where the
http_client
input with an emptypayload
would crash with aurl
containing interpolation functions. - Broker output types (
broker
,try
,switch
) now automatically match the highestmax_in_flight
of their children. The fieldmax_in_flight
can still be manually set in order to enforce a minimum value for when inference isn't possible, such as with dynamic output resources.
- Experimental
azure_renew_lock
field added to theamqp_1
input. - New beta
root_meta
function. - Field
dequeue_visibility_timeout
added to theazure_queue_storage
input. - Field
max_in_flight
added to theazure_queue_storage
output. - New beta Bloblang methods
format_timestamp_unix
andformat_timestamp_unix_nano
. - New Bloblang methods
reverse
andindex_of
. - Experimental
extract_tracing_map
field added to thekafka
input. - Experimental
inject_tracing_map
field added to thekafka
output. - Field
oauth2.scopes
added to HTTP components. - The
mqtt
input and output now support TLS. - Field
enable_renegotiation
added totls
configurations. - Bloblang
if
expressions now support an arbitrary number ofelse if
blocks.
- The
checkpoint_limit
field for thekafka
input now works according to explicit messages in flight rather than the actual offset. This means it now works as expected with compacted topics. - The
aws_kinesis
input should now automatically recover when the shard iterator has expired. - Corrected an issue where messages prefixed with valid JSON documents or values were being decoded in truncated form when the remainder was invalid.
- The following beta components have been promoted to stable:
ristretto
cachecsv
andgenerate
inputsreject
outputbranch
,jq
andworkflow
processors
- Fixed an issue where the
kafka
input with partition balancing wasn't committing offsets.
- The
http_server
input now provides a metadata fieldhttp_server_request_path
. - New methods
sort_by
andkey_values
added to Bloblang.
- Glob patterns for various components no longer resolve to bad paths in the absence of matches.
- Fixed an issue where acknowledgements from the
azure_queue_storage
input would timeout prematurely, resulting in duplicated message delivery. - Unit test definitions no longer have implicit test cases when omitted.
- Vastly improved Bloblang mapping errors.
- The
azure_blob_storage
input will now gracefully terminate if the client credentials become invalid. - Prevented the experimental
gcp_cloud_storage
input from closing early during large file consumption.
- New (experimental) Apache Pulsar input and output.
- Field
codec
added to thesocket
output. - New Bloblang method
map_each_key
. - General config linting improvements.
- Bloblang mappings and interpolated fields within configs are now compile checked during linting.
- New output level
metadata.exclude_prefixes
config field for restricting metadata values sent to the following outputs:kafka
,aws_s3
,amqp_0_9
,redis_streams
,aws_sqs
,gcp_pubsub
. - All NATS components now have
tls
support. - Bloblang now supports context capture in query lambdas.
- New subcommand
benthos blobl server
that hosts a Bloblang editor web application. - New (experimental)
mongodb
output, cache and processor. - New (experimental)
gcp_cloud_storage
input and output. - Field
batch_as_multipart
added to thehttp_client
output. - Inputs, outputs, processors, caches and rate limits now have a component level config field
label
, which sets the metrics and logging prefix. - Resources can now be declared in the new
<component>_resources
fields at the root of config files, the oldresources.<component>s.<label>
style is still valid for backwards compatibility reasons. - Bloblang mappings now support importing the entirety of a map from a path using
from "<path>"
syntax.
- Corrected ack behaviour for the beta
azure_queue_storage
input. - Bloblang compressed arithmetic expressions with field names (
foo+bar
) now correctly parse. - Fixed throughput issues with the
aws_sqs
input. - Prevented using the
root
keyword within Bloblang queries, returning an error message explaining alternative options. Eventuallyroot
references within queries will be fully supported and so returning clear errors messages is a temporary fix. - Increased the offset commit API version used by the
kafka
input to v0.8.2 when consuming explicit partitions.
- Go API: Component implementations now require explicit import from
./public/components/all
in order to be invokable. This should be done automatically at all plugin and custom build entry points. If, however, you notice that your builds have begun complaining that known components do not exist then you will need to explicitly import the package with_ "github.com/Jeffail/benthos/v3/public/components/all"
, if this is the case then please report it as an issue so that it can be dealt with.
- Fixed a potential pipeline stall that would occur when non-batched outputs receive message batches.
- New
azure_queue_storage
input. - All inputs with a
codec
field now support multipart. - New
codec
field added to thehttp_client
,socket
,socket_server
andstdin
inputs. - The
kafka
input now allows an empty consumer group for operating without stored offsets. - The
kafka
input now supports partition ranges.
- The bloblang
encode
method algorithmascii85
no longer returns an error when the input is misaligned.
- The
catch
method now properly executes dynamic argument functions.
- New
http
fieldscert_file
andkey_file
, which when specified enforce HTTPS for the general Benthos server. - Bloblang method
catch
now supportsdeleted()
as an argument.
- Fixed an issue with custom labels becoming stagnant with the
influxdb
metrics type. - Fixed a potential unhandled error when writing to the
azure_queue_storage
output.
- Experimental
sharded_join
fields added to thesequence
input. - Added a new API for writing Bloblang plugins in Go at
./public/bloblang
. - Field
fields_mapping
added to thelog
processor.
- Prevented pre-existing errors from failing/aborting branch execution in the
branch
andworkflow
processors. - Fixed
subprocess
processor message corruption with codecslength_prefixed_uint32_be
andnetstring
.
- The
bloblang
input has been renamed togenerate
. This change is backwards compatible andbloblang
will still be recognized until the next major version release. - Bloblang more often preserves integer precision in arithmetic operations.
- Field
key
in outputredis_list
now supports interpolation functions. - Field
tags
added to outputaws_s3
. - New experimental
sftp
input and output. - New input codec
chunker
. - New field
import_paths
added to theprotobuf
processor, replaces the now deprecatedimport_path
field. - Added format
concatenate
to thearchive
processor.
- The
aws_lambda
processor now adds a metadata fieldlambda_function_error
to messages when the function invocation suffers a runtime error.
- Fixed an issue with the
azure_blob_storage
output whereblob_type
set toAPPEND
could result in send failures. - Fixed a potential panic when shutting down a
socket_server
input with messages in flight. - The
switch
processor now correctly flags errors on messages that cause a check to throw an error.
- New bloblang method
bytes
. - The bloblang method
index
now works on byte arrays. - Field
branch_resources
added to theworkflow
processor. - Field
storage_sas_token
added to theazure_blob_storage
input and output. - The bloblang method
hash
and thehash
processor now supportmd5
. - Field
collector_url
added to thejaeger
tracer. - The bloblang method
strip_html
now allows you to specify a list of allowed elements. - New bloblang method
parse_xml
. - New bloblang method
replace_many
. - New bloblang methods
filepath_split
andfilepath_join
.
- The
cassandra
outputsbackoff.max_elapsed_time
field was unused and has been hidden from docs.
- Field
content_type
andcontent_encoding
added to theamqp_0_9
output. - Batching fields added to the
hdfs
output. - Field
codec_send
andcodec_recv
added to thesubprocess
processor. - Methods
min
,max
,abs
,log
,log10
andceil
added to Bloblang. - Added field
pattern_paths
to thegrok
processor. - The
grok
processor now supports dots within field names for nested values. - New
drop_on
output.
- The
xml
processor now supports non UTF-8 encoding schemes.
- The
drop_on_error
output has been deprecated in favour of the newdrop_on
output.
- New
influxdb
metrics target. - New
azure_blob_storage
input. - New
azure_queue_storage
output. - The
bloblang
input fieldinterval
now supports cron expressions. - New beta
aws_kinesis
andaws_sqs
inputs. - The
bool
bloblang method now supports a wider range of string values. - New
reject
output type for conditionally rejecting messages. - All Redis components now support clustering and fail-over patterns.
- The
compress
anddecompress
processors now support snappy.
- Fixed a panic on startup when using
if
statements within aworkflow
branch request or response map. - The
meta
bloblang function error messages now include the name of the required value. - Config unit tests now report processor errors when checks fail.
- Environment variable interpolations now allow dots within the variable name.
- The experimental
aws_s3
input is now marked as beta. - The beta
kinesis_balanced
input is now deprecated. - All Azure components have been renamed to include the prefix
azure_
, e.g.blob_storage
is nowazure_blob_storage
. The old names can still be used for backwards compatibility. - All AWS components have been renamed to include the prefix
aws_
, e.g.s3
is nowaws_s3
. The old names can still be used for backwards compatibility.
- New field
retry_as_batch
added to thekafka
output to assist in ensuring message ordering through retries. - Field
delay_period
added to the experimentalaws_s3
input. - Added service options for adding API middlewares and specify TLS options for plugin builds.
- Method
not_empty
added to Bloblang. - New
bloblang
predicate type added to unit tests. - Unit test case field
target_processors
now allows you to optionally specify a target file. - Basic auth support added to the
prometheus
metrics pusher.
- Unit tests that define environment variables that are run serially (
parallel: false
) will retain those environment variables during execution, as opposed to only at config parse time. - Lambda distributions now look for config files relative to the binary location, allowing you to deploy configs from the same zip as the binary.
- Add
Content-Type
headers in streams API responses. - Field
delete_objects
is now respected by the experimentalaws_s3
input. - Fixed a case where resource processors couldn't access rate limit resources.
- Input files that are valid according to the codec but empty now trigger acknowledgements.
- Mapping
deleted()
within Bloblang object and array literals now correctly omits the values.
- New field
format
added tologger
supportingjson
andlogfmt
. - The
file
input now provides the metadata fieldpath
on payloads.
- The
output.sent
metric now properly represents the number of individual messages sent even after archiving batches. - Fixed a case where metric processors in streams mode pipelines and dynamic components would hang.
- Sync responses of >1 payloads should now get a correct rfc1341 multipart header.
- The
cassandra
output now correctly marshals float and double values. - The
nanomsg
input with aSUB
socket no longer attempts to set invalid timeout.
- Added field
codec
to thefile
output. - The
file
output now supports dynamic file paths. - Added field
ttl
to thecache
processor and output. - New
sql
output, which is similar to thesql
processor and currently supports Clickhouse, PostgreSQL and MySQL. - The
kafka
input now supports multiple topics, topic partition balancing, and checkpointing. - New
cassandra
output. - Field
allowed_verbs
added to thehttp_server
input and output. - New bloblang function
now
, and methodparse_timestamp
. - New bloblang methods
floor
andround
. - The bloblang method
format_timestamp
now supports strings in ISO 8601 format as well as unix epochs with decimal precision up to nanoseconds.
- The
files
output has been deprecated as its behaviour is now covered byfile
. - The
kafka_balanced
input has now been deprecated as its functionality has been added to thekafka
input. - The
cloudwatch
metrics aggregator is now considered stable. - The
sequence
input is now considered stable. - The
switch
processor no longer permits cases with no processors.
- Fixed the
tar
andtar-gzip
input codecs in experimental inputs. - Fixed a crash that could occur when referencing contextual fields within interpolation functions.
- The
noop
processor can now be inferred with an empty object (noop: {}
). - Fixed potential message corruption with the
file
input when using thelines
codec.
- The
csv
input now supports glob patterns in file paths. - The
file
input now supports multiple paths, glob patterns, and a range of codecs. - New experimental
aws_s3
input. - All
redis
components now support TLS. - The
-r
cli flag now supports glob patterns.
- Bloblang literals, including method and function arguments, can now be mutated without brackets regardless of where they appear.
- Bloblang maps now work when running bloblang with the
blobl
subcommand.
-
The
ristretto
cache no longer forces retries on get commands, and the retry fields have been changed in order to reflect this behaviour. -
The
files
input has been deprecated as its behaviour is now covered byfile
. -
Numbers within JSON documents are now parsed in a way that preserves precision even in cases where the number does not fit a 64-bit signed integer or float. When arithmetic is applied to those numbers (either in Bloblang or by other means) the number is converted (and precision lost) at that point based on the operation itself.
This change means that string coercion on large numbers (e.g.
root.foo = this.large_int.string()
) should now preserve the original form. However, if you are using plugins that interact with JSON message payloads you must ensure that your plugins are able to process thejson.Number
type.This change should otherwise not alter the behaviour of your configs, but if you notice odd side effects you can disable this feature by setting the environment variable
BENTHOS_USE_NUMBER
tofalse
(BENTHOS_USE_NUMBER=false benthos -c ./config.yaml
). Please raise an issue if this is the case so that it can be looked into.
- New input
subprocess
. - New output
subprocess
. - Field
auto_ack
added to theamqp_0_9
input. - Metric labels can be renamed for
prometheus
andcloudwatch
metrics components usingpath_mapping
by assigning meta fields.
- Metrics labels registered using the
rename
metrics component are now sorted before registering, fixing incorrect values that could potentially be seen when renaming multiple metrics to the same name.
- OAuth 2.0 using the client credentials token flow is now supported by the
http_client
input and output, and thehttp
processor. - Method
format_timestamp
added to Bloblang. - Methods
re_find_object
andre_find_all_object
added to Bloblang. - Field
connection_string
added to the Azureblob_storage
andtable_storage
outputs. - Field
public_access_level
added to the Azureblob_storage
output. - Bloblang now supports trailing commas in object and array literals and function and method parameters.
- The
amqp_1
input and output now re-establish connections to brokers on any unknown error. - Batching components now more efficiently attempt a final flush of data during graceful shutdown.
- The
dynamic
output is now more flexible with removing outputs, and should no longer block the API as aggressively.
- New cli flag
log.level
for overriding the configured logging level. - New integration test suite (much more dapper and also a bit more swanky than the last).
- The default value for
batching.count
fields is now zero, which means adding a non-count based batching mechanism without also explicitly overridingcount
no longer incorrectly caps batches at one message. This change is backwards compatible in that working batching configs will not change in behaviour. However, a broken batching config will now behave as expected.
- Improved Bloblang parser error messages for function and method parameters.
- New methods
any
,all
andjson_schema
added to Bloblang. - New function
file
added to Bloblang. - The
switch
output can now route batched messages individually (when using the newcases
field). - The
switch
processor now routes batched messages individually (when using the newcases
field). - The
workflow
processor can now reference resource configuredbranch
processors. - The
metric
processor now has a fieldname
that replaces the now deprecated fieldpath
. When used the processor now applies to all messages of a batch and the name of the metric is now absolute, without being prefixed by a path generated based on its position within the config. - New field
check
added togroup_by
processor children, which now replaces the oldcondition
field. - New field
check
added towhile
processor, which now replaces the oldcondition
field. - New field
check
added toread_until
input, which now replaces the oldcondition
field.
- The
bloblang
input with an interval configured now emits the first message straight away.
- New function
range
added to Bloblang. - New beta
jq
processor. - New driver
clickhouse
added to thesql
processor.
- New field
data_source_name
replacesdsn
for thesql
processor, and when using this field each message of a batch is processed individually. When using the fielddsn
the behaviour remains unchanged for backwards compatibility.
- Eliminated situations where an
amqp_0_9
oramqp_1
component would abandon a connection reset due to partial errors. - The Bloblang parser now allows naked negation of queries.
- The
cache
processor interpolations forkey
andvalue
now cross-batch reference messages before processing.
- New Bloblang methods
not_null
andfilter
. - New Bloblang function
env
. - New field
path_mapping
added to all metrics types. - Field
max_in_flight
added to thedynamic
output. - The
workflow
processor has been updated to usebranch
processors with the new fieldbranches
, these changes are backwards compatible with the now deprecatedstages
field.
- The
rename
,whitelist
andblacklist
metrics types are now deprecated, and thepath_mapping
field should be used instead. - The
conditional
,process_map
andprocess_dag
processors are now deprecated and are superseded by theswitch
,branch
andworkflow
processors respectively.
- Fixed
http
processor error log messages that would print incorrect URLs. - The
http_server
input now emitslatency
metrics. - Fixed a panic that could occur during the shutdown of an
http_server
input serving a backlog of requests. - Explicit component types (
type: foo
) are now checked by the config linter. - The
amqp_1
input and output should now reconnect automatically after an unexpected link detach.
- Improved parser error messages with the
blobl
subcommand. - Added flag
file
to theblobl
subcommand. - New Bloblang method
parse_timestamp_unix
. - New beta
protobuf
processor. - New beta
branch
processor. - Batching fields added to
s3
output.
- The
http
processor fieldmax_parallel
has been deprecated in favour of rate limits, and the fields withinrequest
have been moved to the root of thehttp
namespace. This change is backwards compatible andhttp.request
fields will still be recognized until the next major version release. - The
process_field
processor is now deprecated, andbranch
should be used instead.
- Wholesale metadata mappings (
meta = {"foo":"bar"}
) in Bloblang now correctly clear pre-existing fields.
- Prevented an issue where batched outputs would terminate at start up. Fixes a regression introduced in v3.24.0.
- Endpoint
/ready
added to streams mode API. - Azure
table_storage
output now supports batched sends. - All HTTP components are now able to configure a proxy URL.
- New
ristretto
cache. - Field
shards
added tomemory
cache.
- Batch error handling and retry logic has been improved for the
kafka
anddynamodb
outputs. - Bloblang now allows non-matching not-equals comparisons, allowing
foo != null
expressions.
- Condition
check_interpolation
has been deprecated.
- Path segments in Bloblang mapping targets can now be quote-escaped.
- New beta
sequence
input, for sequentially chaining inputs. - New beta
csv
input for consuming CSV files. - New beta Azure
table_storage
output. - New
parse_csv
Bloblang method. - New
throw
Bloblang function. - The
slice
Bloblang method now supports negative low and high arguments.
- Manual
mqtt
connection handling for both the input and output. This should fix some cases where connections were dropped and never recovered. - Fixed Bloblang error where calls to a
.get
method would returnnull
after the first query. - The
for_each
processor no longer interlaces child processors during split processing.
- Added TLS fields to
elasticsearch
output. - New Bloblang methods
encrypt_aes
anddecrypt_aes
added. - New field
static_headers
added to thekafka
output. - New field
enabled
added to thehttp
config section. - Experimental CLI flag
-resources
added for specifying files containing extra resources.
- The
amqp_0_9
now resolvestype
andkey
fields per message of a batch.
- New beta
bloblang
input for generating documents. - New beta Azure
blob_storage
output. - Field
sync_response.status
added tohttp_server
input. - New Bloblang
errored
function.
- The
json_schema
processor no longer lower cases fields within error messages. - The
dynamodb
cache no longer creates warning logs for get misses.
- SASL config fields added to
amqp_1
input and output. - The
lint
subcommand now supports triple dot wildcard paths:./foo/...
. - The
test
subcommand now supports tests defined within the target config file being tested.
- Bloblang boolean operands now short circuit.
- Fields
strict_mode
andmax_in_flight
added to theswitch
output. - New beta
amqp_1
input and output added.
- Field
drop_empty_bodies
added to thehttp_client
input.
- Fixed deleting and skipping maps with the
blobl
subcommand.
- New field
type
added to theamqp_0_9
output. - New bloblang methods
explode
andwithout
.
- Message functions such as
json
andcontent
now work correctly when executing bloblang with theblobl
sub command.
- New bloblang methods
type
,join
,unique
,escape_html
,unescape_html
,re_find_all
andre_find_all_submatch
. - Bloblang
sort
method now allows custom sorting functions. - Bloblang now supports
if
expressions. - Bloblang now allows joining strings with the
+
operator. - Bloblang now supports multiline strings with triple quotes.
- The
xml
processor is now less strict with XML parsing, allowing unrecognised escape sequences to be passed through unchanged.
- The bloblang method
map_each
now respectsNothing
mapping by copying the underlying value unchanged. - It's now possible to reference resource inputs and outputs in streams mode.
- Fixed a problem with compiling old interpolation functions with arguments containing colons (i.e.
${!timestamp_utc:2006-01-02T15:04:05.000Z}
)
- Flag
log
added totest
sub command to allow logging during tests. - New subcommand
blobl
added for convenient mapping over the command line. - Lots of new bloblang methods.
- The
redis_streams
input no longer incorrectly copies message data into a metadata field.
- Bloblang is no longer considered beta. Therefore, no breaking changes will be introduced outside of a major version release.
- New
ascii85
andz85
options have been added to theencode
anddecode
processors.
- The
meta
function no longer reflects changes made within the map itself. - Extracting data from other messages of a batch using
from
no longer reflects changes made within a map. - Meta assignments are no longer allowed within named maps.
- Assigning
deleted()
toroot
now filters out a message entirely. - Lots of new methods and goodies.
- New HMAC algorithms added to
hash
processor. - New beta
bloblang
processor. - New beta
bloblang
condition.
- Prevented a crash that might occur with high-concurrent access of
http_server
metrics with labels. - The
http_client
output now respects thecopy_response_headers
field.
- Vastly improved function interpolations, including better batch handling and arithmetic operators.
- The
gcp_pubsub
output now supports function interpolation on the fieldtopic
. - New
contains_any
andcontains_any_cs
operators added to thetext
condition. - Support for input and output
resource
types. - The
broker
andswitch
output types now allow async messages and batching within child outputs. - Field
schema_path
added to theavro
processor. - The
redis
cache,redis_list
inputs and outputs now support selecting a database with the URL path. - New field
max_in_flight
added to thebroker
output.
- Benthos now runs in strict mode, but this can be disabled with
--chilled
. - The Benthos CLI has been revamped, the old flags are still supported but are deprecated.
- The
http_server
input now accepts requests without a content-type header.
- Outputs that resolve function interpolations now correctly resolve the
batch_size
function. - The
kinesis_balanced
input now correctly establishes connections. - Fixed an auth transport issue with the
gcp_pubsub
input and output.
- Format
syslog_rfc3164
added to theparse_log
processor. - New
multilevel
cache. - New
json_append
,json_type
andjson_length
functions added to theawk
processor. - New
flatten
operator added to thejson
processor.
- Processors that fail now set the opentracing tag
error
totrue
.
- Kafka connectors now correctly set username and password for all SASL strategies.
- Field
delete_files
added tofiles
input. - TLS fields added to
nsq
input and output. - Field
processors
added to batching fields to easily accommodate aggregations and archiving of batched messages. - New
parse_log
processor. - New
json
condition. - Operators
flatten_array
,fold_number_array
andfold_string_array
added tojson
processor.
- The
redis_streams
input no longer flushes >1 fetched messages as a batch.
- Re-enabled Kafka connections using SASL without TLS.
- New
socket
,socket_server
inputs. - New
socket
output. - Kafka connectors now support SASL using
OAUTHBEARER
,SCRAM-SHA-256
,SCRAM-SHA-512
mechanisms. - Experimental support for AWS CloudWatch metrics.
- The
tcp
,tcp_server
andudp_server
inputs have been deprecated and moved into thesocket
andsocket_server
inputs respectively. - The
udp
andtcp
outputs have been deprecated and moved into thesocket
output.
- The
subprocess
processor now correctly flags errors that occur.
- New field
max_in_flight
added to the following outputs:amqp_0_9
cache
dynamodb
elasticsearch
gcp_pubsub
hdfs
http_client
kafka
kinesis
kinesis_firehose
mqtt
nanomsg
nats
nats_stream
nsq
redis_hash
redis_list
redis_pubsub
redis_streams
s3
sns
sqs
- Batching fields added to the following outputs:
dynamodb
elasticsearch
http_client
kafka
kinesis
kinesis_firehose
sqs
- More TRACE level logs added throughout the pipeline.
- Operator
delete
added tocache
processor. - Operator
explode
added tojson
processor. - Field
storage_class
added tos3
output. - Format
json_map
added tounarchive
processor.
- Function interpolated strings within the
json
processorvalue
field are now correctly unicode escaped. - Retry intervals for
kafka
output have been tuned to prevent circuit breaker throttling.
- New
try
output, which is a drop-in replacement for abroker
with thetry
pattern. - Field
successful_on
added to thehttp
processor. - The
statsd
metrics type now supports Datadog or InfluxDB tagging. - Field
sync_response.headers
added tohttp_server
input. - New
sync_response
processor. - Field
partitioner
added to thekafka
output.
- The
http
processor now gracefully handles empty responses.
- The
kafka
input should now correctly recover from coordinator failures during an offset commit. - Attributes permitted by the
sqs
output should now have parity with real limitations.
- Batching using an input
broker
now works with only one child input configured. - The
zmq4
input now correctly supports broker based batching.
- New
workflow
processor. - New
resource
processor. - Processors can now be registered within the
resources
section of a config.
- The
mqtt
output fieldtopic
field now supports interpolation functions.
- The
kafka
output no longer attempts to send headers on old versions of the protocol.
- New
regexp_expand
operator added to thetext
processor. - New
json_schema
processor.
- New
amqp_0_9
output which replaces the now deprecatedamqp
output. - The
broker
output now supports batching.
- The
memory
buffer now allows parallel processing of batched payloads. - Version and date information should now be correctly displayed in archive distributions.
- The
s3
input now correctly unescapes bucket keys when streaming from SQS.
- Field
sqs_endpoint
added to thes3
input. - Field
kms_key_id
added to thes3
output. - Operator
delete
added tometadata
processor. - New experimental metrics aggregator
stdout
. - Field
ack_wait
added tonats_stream
input. - New
batching
field added tobroker
input for batching merged streams. - Field
healthcheck
added toelasticsearch
output. - New
json_schema
condition.
- Experimental
kafka_cg
input has been removed. - The
kafka_balanced
inputs underlying implementation has been replaced with thekafka_cg
one. - All inputs have been updated to automatically utilise >1 processing threads, with the exception of
kafka
andkinesis
.
- New
is
operator added totext
condition. - New config unit test condition
content_matches
. - Field
init_values
added to thememory
cache. - New
split
operator added tojson
processor. - Fields
user
andpassword
added tomqtt
input and output. - New experimental
amqp_0_9
input.
- Linting is now disabled for the environment var config shipped with docker images, this should prevent the log spam on start up.
- Go API: Experimental
reader.Async
component methods renamed.
- Prevented
kafka_cg
input lock up after batch policy period trigger with no backlog.
- New
redis
processor. - New
kinesis_firehose
output. - New experimental
kafka_cg
input. - Go API: The
metrics.Local
aggregator now supports labels.
- The
json
processor no longer removes content moved from a path to the same path.
This is a major version release, for more information and guidance on how to migrate please refer to https://benthos.dev/docs/guides/migration/v3.
- The
json
processor now allows you tomove
from either a root source or to a root destination. - Added interpolation to the
metadata
processorkey
field. - Granular retry fields added to
kafka
output.
- Go modules are now fully supported, imports must now include the major version (e.g.
github.com/Jeffail/benthos/v3
). - Removed deprecated
mmap_file
buffer. - Removed deprecated (and undocumented) metrics paths.
- Moved field
prefix
from root ofmetrics
into relevant child components. - Names of
process_dag
stages must now match the regexp[a-zA-Z0-9_-]+
. - Go API: buffer constructors now take a
types.Manager
argument in parity with other components. - JSON dot paths within the following components have been updated to allow array-based operations:
awk
processorjson
processorprocess_field
processorprocess_map
processorcheck_field
conditionjson_field
function interpolations3
inputdynamodb
output
- The
sqs
output no longer attempts to send invalid attributes with payloads from metadata. - During graceful shutdown Benthos now scales the attempt to propagate acks for sent messages with the overall system shutdown period.
- The
s3
andsqs
inputs should now correctly log handles and codes from failed SQS message deletes and visibility timeout changes.
- New
message_group_id
andmessage_deduplication_id
fields added tosqs
output for supporting FIFO queues.
- Metadata field
gcp_pubsub_publish_time_unix
added togcp_pubsub
input. - New
tcp
andtcp_server
inputs. - New
udp_server
input. - New
tcp
andudp
outputs. - Metric paths
output.batch.bytes
andoutput.batch.latency
added. - New
rate_limit
processor.
- The
json
processor now correctly stores parsedvalue
JSON when usingset
on the root path.
- The
sqs
input now adds some message attributes as metadata. - Added field
delete_message
tosqs
input. - The
sqs
output now sends metadata as message attributes. - New
batch_policy
field added tomemory
buffer. - New
xml
processor.
- The
prometheus
metrics exporter adds quantiles back to timing metrics.
- Capped slices from lines reader are now enforced.
- The
json
processor now correctly honours anull
value.
- Disabled
kinesis_balanced
input for WASM builds.
- Field
codec
added toprocess_field
processor. - Removed experimental status from sync responses components, which are now considered stable.
- Field
pattern_definitions
added togrok
processor.
- Simplified serverless lambda main function body for improving plugin documentation.
- Fixed a bug where the
prepend
andappend
operators of thetext
processor could result in invalid messages when consuming line-based inputs.
- Field
clean_session
added tomqtt
input. - The
http_server
input now adds request query parameters to messages as metadata.
- Prevent concurrent access race condition on nested parallel
process_map
processors.
- New beta input
kinesis_balanced
. - Field
profile
added to AWS components credentials config.
- Improved error messages attached to payloads that fail
process_dag
. post mappings. - New
redis_hash
output. - New
sns
output.
- Allow extracting metric
rename
submatches into labels. - Field
use_patterns
added toredis_pubsub
input for subscribing to channels using glob-style patterns.
- Go API: It's now possible to specify a custom config unit test file path suffix.
- New rate limit and websocket message fields added to
http_server
input. - The
http
processor now optionally copies headers from response into resulting message metadata. - The
http
processor now sets ahttp_status_code
metadata value into resulting messages (provided one is received.)
- Go API: Removed experimental
Block
functions from the cache and rate limit packages.
- New (experimental) command flags
--test
and--gen-test
added. - All http client components output now set a metric
request_timeout
.
- All errors caught by processors should now be accessible via the
${!error}
interpolation function, rather than just flagged astrue
.
- The
process_field
processor now propagates metadata to the original payload with theresult_type
set to discard. This allows proper error propagation.
- Field
max_buffer
added tosubprocess
processor.
- The
subprocess
processor now correctly logs and recovers subprocess pipeline related errors (such as exceeding buffer limits.)
- New
json_delete
function added to theawk
processor.
- SQS output now correctly waits between retry attempts and escapes error loops during shutdown.
- Go API: Add
RunWithOpts
optOptOverrideConfigDefaults
.
- The
filter
andfilter_parts
config sections now correctly marshall when printing with--all
.
- Go API: A new service method
RunWithOpts
has been added in order to accomodate service customisations with opt funcs.
- New interpolation function
error
.
- New
number
condition. - New
number
processor. - New
avro
processor. - Operator
enum
added totext
condition. - Field
result_type
added toprocess_field
processor for marshalling results into non-string types. - Go API: Plugin APIs now allow nil config constructors.
- Registering plugins automatically adds plugin documentation flags to the main Benthos service.
- Output
http_client
is now able to propagate responses from each request back to inputs supporting sync responses. - Added support for Gzip compression to
http_server
output sync responses. - New
check_interpolation
condition.
- New
sync_response
output type, with experimental support added to thehttp_server
input. - SASL authentication fields added to all Kafka components.
- The
s3
input now setss3_content_encoding
metadata (when not using the download manager.) - New trace logging for the
rename
,blacklist
andwhitelist
metric components to assist with debugging.
- Ability to combine sync and async responses in serverless distributions.
- The
insert_part
,merge_json
andunarchive
processors now propagate message contexts.
- JSON processors no longer escape
&
,<
, and>
characters by default.
- The
http
processor now preserves message metadata and contexts. - Any
http
components that create requests with messages containing empty bodies now correctly function in WASM.
- New
fetch_buffer_cap
field forkafka
andkafka_balanced
inputs. - Input
gcp_pubsub
now has the fieldmax_batch_count
.
- Reduced allocations under most JSON related processors.
- Streams mode API now logs linting errors.
- New interpolation function
batch_size
.
- Output
elasticsearch
no longer reports index not found errors on connect.
- Input reader no longer overrides message contexts for opentracing spans.
- Improved construction error messages for
broker
andswitch
input and outputs.
- Plugins that don't use a configuration structure can now return nil in their sanitise functions in order to have the plugin section omitted.
- The
kafka
andkafka_balanced
inputs now set akafka_lag
metadata field to incoming messages. - The
awk
processor now has a variety of typedjson_set
functionsjson_set_int
,json_set_float
andjson_set_bool
. - Go API: Add experimental function for blocking cache and ratelimit constructors.
- The
json
processor now defaults to an executable operator (clean).
- Add experimental function for blocking processor constructors.
- Core service logic has been moved into new package
service
, making it easier to maintain plugin builds that match upstream Benthos.
- Experimental support for WASM builds.
- Config linting now reports line numbers.
- Config interpolations now support escaping.
- API for creating
cache
implementations. - API for creating
rate_limit
implementations.
This is a major version released due to a series of minor breaking changes, you can read the full migration guide here.
- Benthos now attempts to infer the
type
of config sections whenever the field is omitted, for more information please read this overview: Concise Configuration. - Field
unsubscribe_on_close
of thenats_stream
input is nowfalse
by default.
- The following commandline flags have been removed:
swap-envs
,plugins-dir
,list-input-plugins
,list-output-plugins
,list-processor-plugins
,list-condition-plugins
.
- Package
github.com/Jeffail/benthos/lib/processor/condition
changed togithub.com/Jeffail/benthos/lib/condition
. - Interface
types.Cache
now hastypes.Closable
embedded. - Interface
types.RateLimit
now hastypes.Closable
embedded. - Add method
GetPlugin
to interfacetypes.Manager
. - Add method
WithFields
to interfacelog.Modular
.
- Ensure
process_batch
processor gets normalised correctly.
- New
for_each
processor with the same behaviour asprocess_batch
,process_batch
is now considered an alias forfor_each
.
- The
sql
processor now executes across the batch, documentation updated to clarify.
- Corrected
result_codec
field insql
processor config.
- New
sql
processor.
- Using
json_map_columns
with thedynamodb
output should now correctly storenull
and array values within the target JSON structure.
- New
encode
anddecode
schemehex
.
- Fixed potential panic when attempting an invalid HTTP client configuration.
- Benthos in streams mode no longer tries to load directory
/benthos/streams
by default.
- Field
json_map_columns
added todynamodb
output.
- JSON references are now supported in configuration files.
- The
hash
processor now supportssha1
. - Field
force_path_style_urls
added tos3
components. - Field
content_type
of thes3
output is now interpolated. - Field
content_encoding
added tos3
output.
- The
benthos-lambda
distribution now correctly returns all message parts in synchronous execution.
- Docker builds now use a locally cached
vendor
for dependencies. - All
s3
components no longer default to enforcing path style URLs.
- New output
drop_on_error
. - Field
retry_until_success
added toswitch
output.
- Improved error and output logging for
subprocess
processor when the process exits unexpectedly.
- The main docker image is now based on busybox.
- Lint rule added for
batch
processors outside of the input section.
- Removed potential
benthos-lambda
panic on shut down.
- The
redis
cache no longer incorrectly returns a "key not found" error instead of connection errors.
- Changed docker tag format from
vX.Y.Z
toX.Y.Z
.
- Output
broker
patternfan_out_sequential
. - Output type
drop
for dropping all messages. - New interpolation function
timestamp_utc
.
- New
benthos-lambda
distribution for running Benthos as a lambda function.
- New
s3
cache implementation. - New
file
cache implementation. - Operators
quote
andunquote
added to thetext
processor. - Configs sent via the streams mode HTTP API are now interpolated with environment variable substitutes.
- All AWS
s3
components now enforce path style syntax for bucket URLs. This improves compatibility with third party endpoints.
- New
parallel
processor.
- The
dynamodb
cacheget
call now correctly reports key not found versus general request error.
- New
sqs_bucket_path
field added tos3
input.
- The
sqs
input now rejects messages that fail by resetting the visibility timeout. - The
sqs
input no longer fails to delete consumed messages when the batch contains duplicate message IDs.
- The
metric
processor no longer mixes label keys when processing across parallel pipelines.
- Comma separated
kafka
andkafka_balanced
address and topic values are now trimmed for whitespace.
- Field
max_processing_period
added tokafka
andkafka_balanced
inputs.
- Compaction intervals are now respected by the
memory
cache type.
- Improved
kafka_balanced
consumer group connection behaviour.
- More
kafka_balanced
input config fields for consumer group timeouts.
- New config interpolation function
uuid_v4
.
- The
while
processor now correctly checks conditions against the first batch of the result of last processor loop.
- Field
max_loops
added towhile
processor.
- New
while
processor.
- New
cache
processor. - New
all
condition. - New
any
condition.
- Function interpolation for field
subject
added tonats
output.
- Switched underlying
kafka_balanced
implementation to sarama consumer.
- Always allow acknowledgement flush during graceful termination.
- Removed unnecessary subscription check from
gcp_pubsub
input.
- New field
fields
added tolog
processor for structured log output.
- Function interpolation for field
channel
added toredis_pubsub
output.
- Field
byte_size
added tosplit
processor.
- Field
dependencies
of children of theprocess_dag
processor now correctly parsed from config files.
- Field
push_job_name
added toprometheus
metrics type. - New
rename
metrics target.
- Removed potential race condition in
process_dag
with raw bytes conditions.
- Field
max_batch_count
added tos3
input. - Field
max_number_of_messages
added tosqs
input.
- New
blacklist
metrics target. - New
whitelist
metrics target. - Initial support for opentracing, including a new
tracer
root component. - Improved generated metrics documentation and config examples.
- The
nats_stream
input now has a fieldunsubscribe_on_close
that when disabled allows durable subscription offsets to persist even when all connections are closed. - Metadata field
nats_stream_sequence
added tonats_stream
input.
- The
subprocess
processor no longer sends unexpected empty lines when messages end with a line break.
- New
switch
processor.
- Printing configs now sanitises resource sections.
- The
headers
field inhttp
configs now detects and applieshost
keys.
- New
json_documents
format added to theunarchive
processor. - Field
push_interval
added to theprometheus
metrics type.
- Brokers now correctly parse configs containing plugin types as children.
- Output broker types now correctly allocates nested processors for
fan_out
andtry
patterns. - JSON formatted loggers now correctly escape error messages with line breaks.
- Improved error logging for
s3
input download failures. - More metadata fields copied to messages from the
s3
input. - Field
push_url
added to theprometheus
metrics target.
- Resources (including plugins) that implement
Closable
are now shutdown cleanly.
- New
json_array
format added to thearchive
andunarchive
processors. - Preliminary support added to the resource manager API to allow arbitrary shared resource plugins.
- The
s3
input now caps and iterates batched SQS deletes.
- The
archive
processor now interpolates thepath
per message of the batch.
- Fixed environment variable interpolation when combined with embedded function interpolations.
- Fixed break down metric indexes for input and output brokers.
- Input
s3
can now toggle the use of a download manager, switching off now downloads metadata from the target file. - Output
s3
now writes metadata to the uploaded file. - Operator
unescape_url_query
added totext
processor.
- The
nats_steam
input and output now actively attempt to recover stale connections. - The
awk
processor prints errors and flags failure when the program exits with a non-zero status.
- The
subprocess
processor now attempts to read all flushed stderr output from a process when it fails.
- Function
print_log
added toawk
processor.
- The
awk
processor functionjson_get
no longer returns string values with quotes.
- Processor
awk
codecs changed.
- Output type
sqs
now supports batched message sends.
- Functions
json_get
andjson_set
added toawk
processor.
- Functions
timestamp_format
,timestamp_format_nano
,metadata_get
andmetadata_set
added toawk
processor.
- New
sleep
processor. - New
awk
processor.
- Converted all integer based time period fields to string based, e.g.
timeout_ms: 5000
would now betimeout: 5s
. This will may potentially be disruptive but the--strict
flag should catch all deprecated fields in an existing config.
- Renamed
max_batch_size
tomax_batch_count
for consistency.
- New
max_batch_size
field added tokafka
,kafka_balanced
andamqp
inputs. This provides a mechanism for creating message batches optimistically.
- New
subprocess
processor.
- API: The
types.Processor
interface has been changed in order to add lifetime cleanup methods (addedCloseAsync
andWaitForClose
). For the overwhelming majority of processors these functions will be no-ops. - More consistent
condition
metrics.
- New
try
andcatch
processors for improved processor error handling.
- All processors now attach error flags.
- S3 input is now more flexible with SNS triggered SQS events.
- Processor metrics have been made more consistent.
- New endpoint
/ready
that returns 200 when both the input and output components are connected, otherwise 503. This is intended to be used as a readiness probe.
- Large simplifications to all metrics paths.
- Fully removed the previously deprecated
combine
processor. - Input and output plugins updated to support new connection health checks.
- Field
role_external_id
added to all S3 credential configs. - New
processor_failed
condition and improved processor error handling which can be read about here
- New
content_type
field for thes3
output.
- New
group_by_value
processor.
- Lint errors are logged (level INFO) during normal Benthos operation.
- New
--strict
command flag which causes Benthos to abort when linting errors are found in a config file.
- New
--lint
command flag for linting config files.
- The
s3
output now attempts to batch uploads. - The
s3
input now exposes errors in deleting SQS messages during acks.
- Resource based conditions no longer benefit from cached results. In practice this optimisation was easy to lose in config and difficult to maintain.
- Metadata is now sent to
kafka
outputs. - New
max_inflight
field added to thenats_stream
input.
- Fixed relative path trimming for streams from file directories.
- The
dynamodb
cache and output types now set TTL columns as unix timestamps.
- New
escape_url_query
operator for thetext
processor.
- Removed submatch indexes in the
text
processorfind_regexp
operator and added documentation for expanding submatches in thereplace_regexp
operator.
- Allow submatch indexes in the
find_regexp
operator for thetext
processor.
- New
find_regexp
operator for thetext
processor.
- New
aws
fields to theelasticsearch
output to allow AWS authentication.
- Add max-outstanding fields to
gcp_pubsub
input. - Add new
dynamodb
output.
- The
s3
output now calculatespath
field function interpolations per message of a batch.
- New
set
operator for thetext
processor.
- New
cache
output type.
- New
group_by
processor. - Add bulk send support to
elasticsearch
output.
- New
content
interpolation function.
- New
redis
cache type.
- The
process_map
processor now allows map target path overrides when a target is the parent of another target.
- Field
pipeline
andsniff
added to theelasticsearch
output. - Operators
to_lower
andto_upper
added to thetext
processor.
- Field
endpoint
added to all AWS types.
- Allow
log
config fieldstatic_fields
to be fully overridden.
- New
process_dag
processor. - New
static_fields
map added to log config for setting static log fields.
- JSON log field containing component path moved from
@service
tocomponent
.
- New
gcp_pubsub
input and outputs. - New
log
processor. - New
lambda
processor.
- New
process_batch
processor. - Added
count
field tobatch
processor. - Metrics for
kinesis
output throttles.
- The
combine
processor is now considered DEPRECATED, please use thebatch
processor instead. - The
batch
processor fieldbyte_size
is now set at 0 (and therefore ignored) by default. A log warning has been added in case anyone was relying on the default.
- New
rate_limit
resource with alocal
type. - Field
rate_limit
added tohttp
based processors, inputs and outputs.
- New
prefetch_count
field added tonats
input.
- New
bounds_check
condition type. - New
check_field
condition type. - New
queue
field added tonats
input. - Function interpolation for the
topic
field of thensq
output.
- The
nats
input now defaults to joining a queue.
- The redundant
nsq
output fieldmax_in_flight
has been removed. - The
files
output now interpolates paths per message part of a batch.
- New
hdfs
input and output. - New
switch
output. - New
enum
andhas_prefix
operators for themetadata
condition. - Ability to set
tls
client certificate fields directly.
- New
retry
output. - Added
regex_partial
andregex_exact
operators to themetadata
condition.
- The
kinesis
output fieldretries
has been renamedmax_retries
in order to expose the difference in its zero value behaviour (endless retries) versus otherretry
fields (zero retries).
- New
endpoint
field added tokinesis
input. - New
dynamodb
cache type.
- Function interpolation for the
topic
field of thekafka
output. - New
target_version
field for thekafka_balanced
input. - TLS config fields for client certificates.
- TLS config field
cas_file
has been renamedroot_cas_file
.
- New
zip
option for thearchive
andunarchive
processors.
- The
kinesis
output type now supports batched sends and per message interpolation.
- New
metric
processor.
- New
redis_streams
input and output.
- New
kinesis
input and output.
- The
index
field of theelasticsearch
output can now be dynamically set using function interpolation. - New
hash
processor.
- API: The
metrics.Type
interface has been changed in order to add labels.
- Significant restructuring of
amqp
inputs and outputs. These changes should be backwards compatible for existing pipelines, but changes the way in which queues, exchanges and bindings are declared using these types.
- New durable fields for
amqp
input and output types.
- Improved statsd client with better cached aggregation.
- New
tls
fields foramqp
input and output types.
- New
type
field forelasticsearch
output.
- New
throttle
processor.
- New
less_than
andgreater_than
operators formetadata
condition.
- New
metadata
condition type. - More metadata fields for
kafka
input. - Field
commit_period_ms
forkafka
andkafka_balanced
inputs for specifying a commit period.
- New
retries
field tos3
input, to cap the number of download attempts made on the same bucket item. - Added metadata based mechanism to detect final message from a
read_until
input. - Added field to
split
processor for specifying target batch sizes.
- Metadata fields are now per message part within a batch.
- New
metadata_json_object
function interpolation to return a JSON object of metadata key/value pairs.
- The
metadata
function interpolation now allows part indexing and no longer returns a JSON object when no key is specified, this behaviour can now be done using themetadata_json_object
function.
- Fields for the
http
processor to enable parallel requests from message batches.
- Broker level output processors are now applied before the individual output processors.
- The
dynamic
input and output HTTP paths for CRUD operations are now/inputs/{input_id}
and/outputs/{output_id}
respectively. - Removed deprecated
amazon_s3
,amazon_sqs
andscalability_protocols
input and output types. - Removed deprecated
json_fields
field from thededupe
processor.
- Add conditions to
process_map
processor.
- TLS config fields have been cleaned up for multiple types. This affects the
kafka
,kafka_balanced
andhttp_client
input and output types, as well as thehttp
processor type.
- New
delete_all
anddelete_prefix
operators formetadata
processor. - More metadata fields extracted from the AMQP input.
- HTTP clients now support function interpolation on the URL and header values, this includes the
http_client
input and output as well as thehttp
processor.
- New
key
field added to thededupe
processor, allowing you to deduplicate using function interpolation. This deprecates thejson_paths
array field.
- New
s3
andsqs
input and output types, these replace the now deprecatedamazon_s3
andamazon_sqs
types respectively, which will eventually be removed. - New
nanomsg
input and output types, these replace the now deprecatedscalability_protocols
types, which will eventually be removed.
- Metadata fields are now collected from MQTT input.
- AMQP output writes all metadata as headers.
- AMQP output field
key
now supports function interpolation.
- New
metadata
processor and configuration interpolation function.
- New config interpolator function
json_field
for extracting parts of a JSON message into a config value.
- Log level config field no longer stutters,
logger.log_level
is nowlogger.level
.
- Ability to create batches via conditions on message payloads in the
batch
processor. - New
--examples
flag for generating specific examples from Benthos.
- New
text
processor.
- Processor
process_map
replaced fieldstrict_premapping
withpremap_optional
.
- New
process_field
processor. - New
process_map
processor.
- Removed mapping fields from the
http
processor, this behaviour has been put into the newprocess_map
processor instead.
- Renamed
content
condition type totext
in order to clarify its purpose.
- Latency metrics for caches.
- TLS options for
kafka
andkafka_partitions
inputs and outputs.
- Metrics for items configured within the
resources
section are now namespaced under their identifier.
- New
copy
andmove
operators for thejson
processor.
- Metrics for recording
http
request latencies.
- Improved and rearranged fields for
http_client
input and output.
- More compression and decompression targets.
- New
lines
option for archive/unarchive processors. - New
encode
anddecode
processors. - New
period_ms
field for thebatch
processor. - New
clean
operator for thejson
processor.
- New
http
processor, where payloads can be sent to arbitrary HTTP endpoints and the result constructed into a new payload. - New
inproc
inputs and outputs for linking streams together.
- New streams endpoint
/streams/{id}/stats
for obtaining JSON metrics for a stream.
- Allow comma separated topics for
kafka_balanced
.
- Support for PATCH verb on the streams mode
/streams/{id}
endpoint.
- Sweeping changes were made to the environment variable configuration file. This file is now auto generated along with its supporting document. This change will impact the docker image.
- New
filter_parts
processor for filtering individual parts of a message batch. - New field
open_message
forwebsocket
input.
- No longer setting default input processor.
- New
root_path
field for service widehttp
config.
- New
regexp_exact
andregexp_partial
content condition operators.
- The
statsd
metrics target will now periodically report connection errors.
- The
json
processor will nowappend
array values in expanded form.
- More granular config options in the
http_client
output for controlling retry logic. - New
try
pattern for the outputbroker
type, which can be used in order to configure fallback outputs. - New
json
processor, this replacesdelete_json
,select_json
,set_json
.
- The
streams
API endpoints have been changed to become more "RESTy". - Removed the
delete_json
,select_json
andset_json
processors, please use thejson
processor instead.
- New
grok
processor for creating structured objects from unstructured data.
- New
files
input type for reading multiple files as discrete messages.
- Increase default
max_buffer
forstdin
,file
andhttp_client
inputs. - Command flags
--print-yaml
and--print-json
changed to provide sanitised outputs unless accompanied by new--all
flag.
- Badger based buffer option has been removed.
- New metrics wrapper for more basic interface implementations.
- New
delete_json
processor. - New field
else_processors
forconditional
processor.
- New websocket endpoint for
http_server
input. - New websocket endpoint for
http_server
output. - New
websocket
input type. - New
websocket
output type.
- Goreleaser config for generating release packages.
- Back to using Scratch as base for Docker image, instead taking ca-certificates from the build image.
- New
batch
processor for combining payloads up to a number of bytes. - New
conditional
processor, allows you to configure a chain of processors to only be run if the payload passes acondition
. - New
--stream
mode features:- POST verb for
/streams
path now supported. - New
--streams-dir
flag for parsing a directory of stream configs.
- POST verb for
- The
condition
processor has been renamedfilter
. - The
custom_delimiter
fields in any line reader typesfile
,stdin
,stdout
, etc have been renameddelimiter
, where the behaviour is the same. - Now using Alpine as base for Docker image, includes ca-certificates.