PROXY, a lightweight elasticsearch proxy written in golang.
- Auto handling upstream failure while indexing, aka nonstop indexing
- Auto detect the upstream failure in search
- Multiple write mechanism, one indexing request map to multi remote elasticsearch clusters
- Support TLS/HTTPS, generate the cert files automatically
- Support run background as daemon mode(only available on linux and macOS)
- Auto merge indexing operations to single bulk operation(WIP)
- Load balancing(indexing and search request), algorithm configurable(WIP)
- A controllable query cache layer, use redis as backend
- Index throttling or buffering, via disk based indexing queue(limit by queue length or size)
- Search throttling, limit concurrent connections to upstream(WIP)
- Builtin stats API and management UI(WIP)
- Builtin floating IP, support seamless failover and rolling upgrade
- First, setup upstream config in the
proxy.yml
.
api:
enabled: true
network:
binding: 0.0.0.0:2900
tls:
enabled: true
elasticsearch:
- name: default
enabled: true
endpoint: http://localhost:9200
plugins:
- name: proxy
enabled: true
upstream:
- name: primary
enabled: true
rate_limit:
max_qps: 10000
queue_name: primary
max_queue_depth: -1
timeout: 60s
elasticsearch: default
- Start the PROXY.
➜ elasticsearch-proxy ✗ ./bin/proxy
___ ____ ____ _ _ _ _
|__] |__/ | | \/ \_/
| | \ |__| _/\_ |
[PROXY] An elasticsearch proxy written in golang.
0.1.0_SNAPSHOT, 430bd60, Sun Apr 8 09:44:38 2018 +0800, medcl, seems good to go
[04-05 19:30:13] [INF] [instance.go:23] workspace: data/APP/nodes/0
[04-05 19:30:13] [INF] [api.go:147] api server listen at: https://0.0.0.0:2900
- Done! Now you are ready to rock with it.
➜ curl -k -XGET https://localhost:2900/
{
"name": "PROXY",
"tagline": "You Know, for Proxy",
"upstream": {
"backup": "http://localhost:9201",
"primary": "http://localhost:9200"
},
"uptime": "1m58.019165s",
"version": {
"build_commit": "430bd60, Sun Apr 8 09:44:38 2018 +0800, medcl, seems good to go ",
"build_date": "Sun Apr 8 09:58:29 CST 2018",
"number": "0.1.0_SNAPSHOT"
}
}
➜ curl -k -XGET -H'UPSTREAM:primary' https://localhost:2900/
{
"name" : "XZDZ8qc",
"cluster_name" : "my-application",
"cluster_uuid" : "FWt_UO6BRr6uBVhkVrisew",
"version" : {
"number" : "6.2.3",
"build_hash" : "c59ff00",
"build_date" : "2018-03-13T10:06:29.741383Z",
"build_snapshot" : false,
"lucene_version" : "7.2.1",
"minimum_wire_compatibility_version" : "5.6.0",
"minimum_index_compatibility_version" : "5.0.0"
},
"tagline" : "You Know, for Search"
}
➜ curl -k -XGET -H'UPSTREAM:backup' https://localhost:2900/
{
"name" : "zRcp1My",
"cluster_name" : "elasticsearch",
"cluster_uuid" : "FWt_UO6BRr6uBVhkVrisew",
"version" : {
"number" : "5.6.8",
"build_hash" : "688ecce",
"build_date" : "2018-02-16T16:46:30.010Z",
"build_snapshot" : false,
"lucene_version" : "6.6.1"
},
"tagline" : "You Know, for Search"
}
➜ curl -k -XPOST https://localhost:2900/myindex/_doc/1 -d'{"msg":"hello world!"}'
{ "acknowledge": true }
➜ curl -k -XGET https://localhost:2900/myindex/_doc/1
{"_index":"myindex","_type":"_doc","_id":"1","_version":1,"found":true,"_source":{"msg":"hello world!"}}
➜ curl -k -XPUT https://localhost:2900/myindex/_doc/1 -d'{"msg":"i am a proxy!"}'
{ "acknowledge": true }
➜ curl -k -XGET https://localhost:2900/myindex/_doc/1
{"_index":"myindex","_type":"_doc","_id":"1","_version":2,"found":true,"_source":{"msg":"i am a proxy!"}}
➜ curl -k -XGET https://localhost:2900/myindex/_search?q=proxy
{"took":171,"timed_out":false,"_shards":{"total":1,"successful":1,"skipped":0,"failed":0},"hits":{"total":{"value":1,"relation":"eq"},"max_score":0.8547784,"hits":[{"_index":"myindex","_type":"_doc","_id":"1","_score":0.8547784,"_source":{"msg":"i am a proxy!"}}]}}
➜ curl -k -XDELETE https://localhost:2900/myindex/_doc/1
{ "acknowledge": true }
➜ curl -k -XGET https://localhost:2900/myindex/_doc/1
{"_index":"myindex","_type":"_doc","_id":"1","found":false}
Have fun!
-
Additional request headers
UPSTREAM
, manually choose which upstream are going to query against(read/search requests)
➜ curl -v -XGET -H'UPSTREAM:primary' https://localhost:2900/index/_doc/1 Note: Unnecessary use of -X or --request, GET is already inferred. * Trying 127.0.0.1... * TCP_NODELAY set * Connected to localhost (127.0.0.1) port 2900 (#0) > GET /index/_doc/1 HTTP/1.1 > Host: localhost:2900 > User-Agent: curl/7.54.0 > Accept: */* > UPSTREAM:primary > < HTTP/1.1 200 OK < Upstream: primary < Date: Sat, 07 Apr 2018 13:00:30 GMT < Content-Length: 86 < Content-Type: text/plain; charset=utf-8 < * Connection #0 to host localhost left intact {"_index":"index","_type":"_doc","_id":"1","_version":5,"found":true,"_source":{"a":6}}%
In order to use floating IP, configure floating_ip
under plugin section, set ip
to a unoccupied ip which will be used to share across proxies, set interface
to which network device will be used to bind floating ip.
plugins:
- name: floating_ip
enabled: true
ip: 192.168.1.222
netmask: 255.255.255.0
interface: en0
priority: 100
Note: Floating IP feature may not support on docker/container platform, and should not deploy multi proxy instances on single host.
- Status
curl -k -XGET https://localhost:2900/_proxy/stats
curl -k -XGET https://localhost:2900/_proxy/queue/stats
- Resume Queue
curl -k -XPOST https://localhost:2900/_proxy/queue/resume -d'{"queue":"primary"}'
- Get Error requests
curl -k -XGET https://localhost:2900/_proxy/requests/?from=0&size=20&upstream=primary&status=1
- Replay Error log
curl -k -XPOST https://localhost:2900/_proxy/request/redo -d'{"ids":["bb6t4cqaukihf1ag10q0","bb6t4daaukihf1ag10r0"]}'
MacBook Pro (13-inch, 2017, Four Thunderbolt 3 Ports), 3.5 GHz Intel Core i7, 16 GB 2133 MHz LPDDR3
- Https 2900, query
~$ wrk -c 1000 -d 3m -t 10 -H --latency https://localhost:2900/index/_search?q=customer:A
Running 3m test @ https://localhost:2900/index/_search?q=customer:A
10 threads and 1000 connections
Thread Stats Avg Stdev Max +/- Stdev
Latency 19.13ms 9.07ms 247.10ms 75.79%
Req/Sec 2.63k 699.58 10.55k 61.93%
4707681 requests in 3.00m, 1.89GB read
Socket errors: connect 417, read 290, write 0, timeout 0
Requests/sec: 26140.53
Transfer/sec: 10.72MB
- Https 2900, query with cache
~$ wrk -c 1000 -d 3m -t 10 -H --latency https://localhost:2900/index/_search?q=customer:A
Running 3m test @ https://localhost:2900/index/_search?q=customer:A
10 threads and 1000 connections
Thread Stats Avg Stdev Max +/- Stdev
Latency 7.59ms 8.47ms 338.09ms 87.25%
Req/Sec 6.88k 3.08k 38.47k 78.65%
12275137 requests in 3.00m, 4.90GB read
Socket errors: connect 387, read 676, write 0, timeout 0
Requests/sec: 68158.61
Transfer/sec: 27.84MB
go1.14+
mkdir ~/go/src/infini.sh/ -p
cd ~/go/src/infini.sh/
git clone https://github.com/medcl/elasticsearch-proxy.git proxy
cd proxy
make
Note: Path matters, please make sure follow exactly the above steps.
The docker image size is only 8.7 MB.
Pull it from official docker hub
docker pull medcl/elasticsearch-proxy:latest
Or build your own image locally
docker build -t medcl/elasticsearch-proxy:latest -f docker/Dockerfile .
Customize your proxy.yml
, place somewhere, eg: /tmp/proxy.yml
tee /tmp/proxy.yml <<-'EOF'
elasticsearch:
- name: default
enabled: true
endpoint: http://192.168.3.123:9200
index_prefix: proxy-
basic_auth:
username: elastic
password: changeme
EOF
Rock with your proxy!
docker run --publish 2900:2900 -v /tmp/proxy.yml:/proxy.yml medcl/elasticsearch-proxy:latest
Released under the Apache License, Version 2.0.