2019-09-09 21:23:27 +03:00
|
|
|
// Copyright 2015 Matthew Holt and The Caddy Authors
|
|
|
|
//
|
|
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
// you may not use this file except in compliance with the License.
|
|
|
|
// You may obtain a copy of the License at
|
|
|
|
//
|
|
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
//
|
|
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
// See the License for the specific language governing permissions and
|
|
|
|
// limitations under the License.
|
|
|
|
|
|
|
|
package reverseproxy
|
|
|
|
|
|
|
|
import (
|
2021-01-04 21:11:36 +03:00
|
|
|
"log"
|
2020-02-28 06:56:24 +03:00
|
|
|
"net"
|
2019-09-09 21:23:27 +03:00
|
|
|
"net/http"
|
2020-02-28 06:56:24 +03:00
|
|
|
"reflect"
|
2019-09-09 21:23:27 +03:00
|
|
|
"strconv"
|
|
|
|
"strings"
|
|
|
|
|
|
|
|
"github.com/caddyserver/caddy/v2"
|
|
|
|
"github.com/caddyserver/caddy/v2/caddyconfig"
|
|
|
|
"github.com/caddyserver/caddy/v2/caddyconfig/caddyfile"
|
|
|
|
"github.com/caddyserver/caddy/v2/caddyconfig/httpcaddyfile"
|
|
|
|
"github.com/caddyserver/caddy/v2/modules/caddyhttp"
|
2019-09-20 22:13:49 +03:00
|
|
|
"github.com/caddyserver/caddy/v2/modules/caddyhttp/headers"
|
2019-09-09 21:23:27 +03:00
|
|
|
"github.com/dustin/go-humanize"
|
|
|
|
)
|
|
|
|
|
|
|
|
func init() {
|
|
|
|
httpcaddyfile.RegisterHandlerDirective("reverse_proxy", parseCaddyfile)
|
reverseproxy: copy_response and copy_response_headers for handle_response routes (#4391)
* reverseproxy: New `copy_response` handler for `handle_response` routes
Followup to #4298 and #4388.
This adds a new `copy_response` handler which may only be used in `reverse_proxy`'s `handle_response` routes, which can be used to actually copy the proxy response downstream.
Previously, if `handle_response` was used (with routes, not the status code mode), it was impossible to use the upstream's response body at all, because we would always close the body, expecting the routes to write a new body from scratch.
To implement this, I had to refactor `h.reverseProxy()` to move all the code that came after the `HandleResponse` loop into a new function. This new function `h.finalizeResponse()` takes care of preparing the response by removing extra headers, dealing with trailers, then copying the headers and body downstream.
Since basically what we want `copy_response` to do is invoke `h.finalizeResponse()` at a configurable point in time, we need to pass down the proxy handler, the response, and some other state via a new `req.WithContext(ctx)`. Wrapping a new context is pretty much the only way we have to jump a few layers in the HTTP middleware chain and let a handler pick up this information. Feels a bit dirty, but it works.
Also fixed a bug with the `http.reverse_proxy.upstream.duration` placeholder, it always had the same duration as `http.reverse_proxy.upstream.latency`, but the former was meant to be the time taken for the roundtrip _plus_ copying/writing the response.
* Delete the "Content-Length" header if we aren't copying
Fixes a bug where the Content-Length will mismatch the actual bytes written if we skipped copying the response, so we get a message like this when using curl:
```
curl: (18) transfer closed with 18 bytes remaining to read
```
To replicate:
```
{
admin off
debug
}
:8881 {
reverse_proxy 127.0.0.1:8882 {
@200 status 200
handle_response @200 {
header Foo bar
}
}
}
:8882 {
header Content-Type application/json
respond `{"hello": "world"}` 200
}
```
* Implement `copy_response_headers`, with include/exclude list support
* Apply suggestions from code review
Co-authored-by: Matt Holt <mholt@users.noreply.github.com>
2022-03-09 21:00:51 +03:00
|
|
|
httpcaddyfile.RegisterHandlerDirective("copy_response", parseCopyResponseCaddyfile)
|
|
|
|
httpcaddyfile.RegisterHandlerDirective("copy_response_headers", parseCopyResponseHeadersCaddyfile)
|
2019-09-09 21:23:27 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
func parseCaddyfile(h httpcaddyfile.Helper) (caddyhttp.MiddlewareHandler, error) {
|
|
|
|
rp := new(Handler)
|
|
|
|
err := rp.UnmarshalCaddyfile(h.Dispenser)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
2021-05-02 21:39:06 +03:00
|
|
|
err = rp.FinalizeUnmarshalCaddyfile(h)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
2019-09-09 21:23:27 +03:00
|
|
|
return rp, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// UnmarshalCaddyfile sets up the handler from Caddyfile tokens. Syntax:
|
|
|
|
//
|
|
|
|
// reverse_proxy [<matcher>] [<upstreams...>] {
|
2022-03-13 09:38:11 +03:00
|
|
|
// # backends
|
|
|
|
// to <upstreams...>
|
2022-03-07 03:43:39 +03:00
|
|
|
// dynamic <name> [...]
|
2019-09-09 21:23:27 +03:00
|
|
|
//
|
|
|
|
// # load balancing
|
|
|
|
// lb_policy <name> [<options...>]
|
|
|
|
// lb_try_duration <duration>
|
|
|
|
// lb_try_interval <interval>
|
|
|
|
//
|
|
|
|
// # active health checking
|
2022-03-13 09:38:11 +03:00
|
|
|
// health_uri <uri>
|
|
|
|
// health_port <port>
|
2019-09-09 21:23:27 +03:00
|
|
|
// health_interval <interval>
|
2022-03-13 09:38:11 +03:00
|
|
|
// health_timeout <duration>
|
|
|
|
// health_status <status>
|
|
|
|
// health_body <regexp>
|
2021-01-04 21:26:18 +03:00
|
|
|
// health_headers {
|
|
|
|
// <field> [<values...>]
|
|
|
|
// }
|
2019-09-09 21:23:27 +03:00
|
|
|
//
|
|
|
|
// # passive health checking
|
2022-03-13 09:38:11 +03:00
|
|
|
// fail_duration <duration>
|
|
|
|
// max_fails <num>
|
|
|
|
// unhealthy_status <status>
|
2019-09-09 21:23:27 +03:00
|
|
|
// unhealthy_latency <duration>
|
2022-03-13 09:38:11 +03:00
|
|
|
// unhealthy_request_count <num>
|
2019-09-09 21:23:27 +03:00
|
|
|
//
|
2019-11-27 21:51:32 +03:00
|
|
|
// # streaming
|
|
|
|
// flush_interval <duration>
|
2020-10-30 21:05:21 +03:00
|
|
|
// buffer_requests
|
2022-03-13 09:38:11 +03:00
|
|
|
// buffer_responses
|
|
|
|
// max_buffer_size <size>
|
2019-11-27 21:51:32 +03:00
|
|
|
//
|
2019-09-20 22:13:49 +03:00
|
|
|
// # header manipulation
|
2022-03-07 02:51:55 +03:00
|
|
|
// trusted_proxies [private_ranges] <ranges...>
|
2019-09-20 22:13:49 +03:00
|
|
|
// header_up [+|-]<field> [<value|regexp> [<replacement>]]
|
|
|
|
// header_down [+|-]<field> [<value|regexp> [<replacement>]]
|
|
|
|
//
|
2019-09-09 21:23:27 +03:00
|
|
|
// # round trip
|
|
|
|
// transport <name> {
|
|
|
|
// ...
|
|
|
|
// }
|
2021-05-02 21:39:06 +03:00
|
|
|
//
|
2022-03-13 09:38:11 +03:00
|
|
|
// # optionally intercept responses from upstream
|
2021-05-02 21:39:06 +03:00
|
|
|
// @name {
|
|
|
|
// status <code...>
|
|
|
|
// header <field> [<value>]
|
|
|
|
// }
|
2022-03-13 09:38:11 +03:00
|
|
|
// replace_status [<matcher>] <status_code>
|
2022-03-02 00:12:43 +03:00
|
|
|
// handle_response [<matcher>] {
|
2021-05-02 21:39:06 +03:00
|
|
|
// <directives...>
|
2022-03-13 09:38:11 +03:00
|
|
|
//
|
|
|
|
// # special directives only available in handle_response
|
|
|
|
// copy_response [<matcher>] [<status>] {
|
|
|
|
// status <status>
|
|
|
|
// }
|
|
|
|
// copy_response_headers [<matcher>] {
|
|
|
|
// include <fields...>
|
|
|
|
// exclude <fields...>
|
|
|
|
// }
|
2021-05-02 21:39:06 +03:00
|
|
|
// }
|
2019-09-09 21:23:27 +03:00
|
|
|
// }
|
|
|
|
//
|
2020-02-28 06:56:24 +03:00
|
|
|
// Proxy upstream addresses should be network dial addresses such
|
|
|
|
// as `host:port`, or a URL such as `scheme://host:port`. Scheme
|
|
|
|
// and port may be inferred from other parts of the address/URL; if
|
|
|
|
// either are missing, defaults to HTTP.
|
2021-05-02 21:39:06 +03:00
|
|
|
//
|
|
|
|
// The FinalizeUnmarshalCaddyfile method should be called after this
|
|
|
|
// to finalize parsing of "handle_response" blocks, if possible.
|
2019-09-09 21:23:27 +03:00
|
|
|
func (h *Handler) UnmarshalCaddyfile(d *caddyfile.Dispenser) error {
|
2020-02-28 06:56:24 +03:00
|
|
|
// currently, all backends must use the same scheme/protocol (the
|
|
|
|
// underlying JSON does not yet support per-backend transports)
|
|
|
|
var commonScheme string
|
|
|
|
|
|
|
|
// we'll wait until the very end of parsing before
|
|
|
|
// validating and encoding the transport
|
|
|
|
var transport http.RoundTripper
|
|
|
|
var transportModuleName string
|
|
|
|
|
2021-05-02 21:39:06 +03:00
|
|
|
// collect the response matchers defined as subdirectives
|
|
|
|
// prefixed with "@" for use with "handle_response" blocks
|
|
|
|
h.responseMatchers = make(map[string]caddyhttp.ResponseMatcher)
|
|
|
|
|
2020-03-24 19:53:53 +03:00
|
|
|
// appendUpstream creates an upstream for address and adds
|
|
|
|
// it to the list. If the address starts with "srv+" it is
|
|
|
|
// treated as a SRV-based upstream, and any port will be
|
|
|
|
// dropped.
|
|
|
|
appendUpstream := func(address string) error {
|
|
|
|
isSRV := strings.HasPrefix(address, "srv+")
|
|
|
|
if isSRV {
|
|
|
|
address = strings.TrimPrefix(address, "srv+")
|
|
|
|
}
|
2022-03-06 02:34:19 +03:00
|
|
|
|
|
|
|
dialAddr, scheme, err := parseUpstreamDialAddress(address)
|
2020-03-24 19:53:53 +03:00
|
|
|
if err != nil {
|
2022-03-06 02:34:19 +03:00
|
|
|
return d.WrapErr(err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// the underlying JSON does not yet support different
|
|
|
|
// transports (protocols or schemes) to each backend,
|
|
|
|
// so we remember the last one we see and compare them
|
|
|
|
if commonScheme != "" && scheme != commonScheme {
|
|
|
|
return d.Errf("for now, all proxy upstreams must use the same scheme (transport protocol); expecting '%s://' but got '%s://'",
|
|
|
|
commonScheme, scheme)
|
2020-03-24 19:53:53 +03:00
|
|
|
}
|
2022-03-06 02:34:19 +03:00
|
|
|
commonScheme = scheme
|
|
|
|
|
2020-03-24 19:53:53 +03:00
|
|
|
if isSRV {
|
|
|
|
if host, _, err := net.SplitHostPort(dialAddr); err == nil {
|
|
|
|
dialAddr = host
|
|
|
|
}
|
|
|
|
h.Upstreams = append(h.Upstreams, &Upstream{LookupSRV: dialAddr})
|
|
|
|
} else {
|
|
|
|
h.Upstreams = append(h.Upstreams, &Upstream{Dial: dialAddr})
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2019-09-09 21:23:27 +03:00
|
|
|
for d.Next() {
|
|
|
|
for _, up := range d.RemainingArgs() {
|
2020-03-24 19:53:53 +03:00
|
|
|
err := appendUpstream(up)
|
2020-02-28 06:56:24 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2019-09-09 21:23:27 +03:00
|
|
|
}
|
|
|
|
|
2019-09-11 04:21:52 +03:00
|
|
|
for d.NextBlock(0) {
|
2021-05-02 21:39:06 +03:00
|
|
|
// if the subdirective has an "@" prefix then we
|
|
|
|
// parse it as a response matcher for use with "handle_response"
|
|
|
|
if strings.HasPrefix(d.Val(), matcherPrefix) {
|
|
|
|
err := caddyhttp.ParseNamedResponseMatcher(d.NewFromNextSegment(), h.responseMatchers)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
2019-09-09 21:23:27 +03:00
|
|
|
switch d.Val() {
|
|
|
|
case "to":
|
|
|
|
args := d.RemainingArgs()
|
|
|
|
if len(args) == 0 {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
for _, up := range args {
|
2020-03-24 19:53:53 +03:00
|
|
|
err := appendUpstream(up)
|
2020-02-28 06:56:24 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2019-09-09 21:23:27 +03:00
|
|
|
}
|
|
|
|
|
2022-03-07 03:43:39 +03:00
|
|
|
case "dynamic":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if h.DynamicUpstreams != nil {
|
|
|
|
return d.Err("dynamic upstreams already specified")
|
|
|
|
}
|
|
|
|
dynModule := d.Val()
|
|
|
|
modID := "http.reverse_proxy.upstreams." + dynModule
|
|
|
|
unm, err := caddyfile.UnmarshalModule(d, modID)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
source, ok := unm.(UpstreamSource)
|
|
|
|
if !ok {
|
|
|
|
return d.Errf("module %s (%T) is not an UpstreamSource", modID, unm)
|
|
|
|
}
|
|
|
|
h.DynamicUpstreamsRaw = caddyconfig.JSONModuleObject(source, "source", dynModule, nil)
|
|
|
|
|
2019-09-09 21:23:27 +03:00
|
|
|
case "lb_policy":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if h.LoadBalancing != nil && h.LoadBalancing.SelectionPolicyRaw != nil {
|
|
|
|
return d.Err("load balancing selection policy already specified")
|
|
|
|
}
|
|
|
|
name := d.Val()
|
2021-01-06 00:39:30 +03:00
|
|
|
modID := "http.reverse_proxy.selection_policies." + name
|
|
|
|
unm, err := caddyfile.UnmarshalModule(d, modID)
|
2019-09-09 21:23:27 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
sel, ok := unm.(Selector)
|
|
|
|
if !ok {
|
2021-01-06 00:39:30 +03:00
|
|
|
return d.Errf("module %s (%T) is not a reverseproxy.Selector", modID, unm)
|
2019-09-09 21:23:27 +03:00
|
|
|
}
|
|
|
|
if h.LoadBalancing == nil {
|
|
|
|
h.LoadBalancing = new(LoadBalancing)
|
|
|
|
}
|
|
|
|
h.LoadBalancing.SelectionPolicyRaw = caddyconfig.JSONModuleObject(sel, "policy", name, nil)
|
|
|
|
|
|
|
|
case "lb_try_duration":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if h.LoadBalancing == nil {
|
|
|
|
h.LoadBalancing = new(LoadBalancing)
|
|
|
|
}
|
2020-05-12 01:41:11 +03:00
|
|
|
dur, err := caddy.ParseDuration(d.Val())
|
2019-09-09 21:23:27 +03:00
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad duration value %s: %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
h.LoadBalancing.TryDuration = caddy.Duration(dur)
|
|
|
|
|
|
|
|
case "lb_try_interval":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if h.LoadBalancing == nil {
|
|
|
|
h.LoadBalancing = new(LoadBalancing)
|
|
|
|
}
|
2020-05-12 01:41:11 +03:00
|
|
|
dur, err := caddy.ParseDuration(d.Val())
|
2019-09-09 21:23:27 +03:00
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad interval value '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
h.LoadBalancing.TryInterval = caddy.Duration(dur)
|
|
|
|
|
2021-03-30 03:36:40 +03:00
|
|
|
case "health_uri":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if h.HealthChecks == nil {
|
|
|
|
h.HealthChecks = new(HealthChecks)
|
|
|
|
}
|
|
|
|
if h.HealthChecks.Active == nil {
|
|
|
|
h.HealthChecks.Active = new(ActiveHealthChecks)
|
|
|
|
}
|
|
|
|
h.HealthChecks.Active.URI = d.Val()
|
|
|
|
|
2019-09-09 21:23:27 +03:00
|
|
|
case "health_path":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if h.HealthChecks == nil {
|
|
|
|
h.HealthChecks = new(HealthChecks)
|
|
|
|
}
|
|
|
|
if h.HealthChecks.Active == nil {
|
|
|
|
h.HealthChecks.Active = new(ActiveHealthChecks)
|
|
|
|
}
|
|
|
|
h.HealthChecks.Active.Path = d.Val()
|
2021-03-30 03:36:40 +03:00
|
|
|
caddy.Log().Named("config.adapter.caddyfile").Warn("the 'health_path' subdirective is deprecated, please use 'health_uri' instead!")
|
2019-09-09 21:23:27 +03:00
|
|
|
|
|
|
|
case "health_port":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if h.HealthChecks == nil {
|
|
|
|
h.HealthChecks = new(HealthChecks)
|
|
|
|
}
|
|
|
|
if h.HealthChecks.Active == nil {
|
|
|
|
h.HealthChecks.Active = new(ActiveHealthChecks)
|
|
|
|
}
|
|
|
|
portNum, err := strconv.Atoi(d.Val())
|
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad port number '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
h.HealthChecks.Active.Port = portNum
|
|
|
|
|
2021-01-04 21:26:18 +03:00
|
|
|
case "health_headers":
|
|
|
|
healthHeaders := make(http.Header)
|
2021-12-17 18:53:11 +03:00
|
|
|
for nesting := d.Nesting(); d.NextBlock(nesting); {
|
|
|
|
key := d.Val()
|
|
|
|
values := d.RemainingArgs()
|
|
|
|
if len(values) == 0 {
|
|
|
|
values = append(values, "")
|
2021-01-04 21:26:18 +03:00
|
|
|
}
|
2021-12-17 18:53:11 +03:00
|
|
|
healthHeaders[key] = values
|
2021-01-04 21:26:18 +03:00
|
|
|
}
|
|
|
|
if h.HealthChecks == nil {
|
|
|
|
h.HealthChecks = new(HealthChecks)
|
|
|
|
}
|
|
|
|
if h.HealthChecks.Active == nil {
|
|
|
|
h.HealthChecks.Active = new(ActiveHealthChecks)
|
|
|
|
}
|
|
|
|
h.HealthChecks.Active.Headers = healthHeaders
|
|
|
|
|
2019-09-09 21:23:27 +03:00
|
|
|
case "health_interval":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if h.HealthChecks == nil {
|
|
|
|
h.HealthChecks = new(HealthChecks)
|
|
|
|
}
|
|
|
|
if h.HealthChecks.Active == nil {
|
|
|
|
h.HealthChecks.Active = new(ActiveHealthChecks)
|
|
|
|
}
|
2020-05-12 01:41:11 +03:00
|
|
|
dur, err := caddy.ParseDuration(d.Val())
|
2019-09-09 21:23:27 +03:00
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad interval value %s: %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
h.HealthChecks.Active.Interval = caddy.Duration(dur)
|
|
|
|
|
|
|
|
case "health_timeout":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if h.HealthChecks == nil {
|
|
|
|
h.HealthChecks = new(HealthChecks)
|
|
|
|
}
|
|
|
|
if h.HealthChecks.Active == nil {
|
|
|
|
h.HealthChecks.Active = new(ActiveHealthChecks)
|
|
|
|
}
|
2020-05-12 01:41:11 +03:00
|
|
|
dur, err := caddy.ParseDuration(d.Val())
|
2019-09-09 21:23:27 +03:00
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad timeout value %s: %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
h.HealthChecks.Active.Timeout = caddy.Duration(dur)
|
|
|
|
|
|
|
|
case "health_status":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if h.HealthChecks == nil {
|
|
|
|
h.HealthChecks = new(HealthChecks)
|
|
|
|
}
|
|
|
|
if h.HealthChecks.Active == nil {
|
|
|
|
h.HealthChecks.Active = new(ActiveHealthChecks)
|
|
|
|
}
|
|
|
|
val := d.Val()
|
|
|
|
if len(val) == 3 && strings.HasSuffix(val, "xx") {
|
|
|
|
val = val[:1]
|
|
|
|
}
|
2021-03-30 03:36:40 +03:00
|
|
|
statusNum, err := strconv.Atoi(val)
|
2019-09-09 21:23:27 +03:00
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad status value '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
h.HealthChecks.Active.ExpectStatus = statusNum
|
|
|
|
|
|
|
|
case "health_body":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if h.HealthChecks == nil {
|
|
|
|
h.HealthChecks = new(HealthChecks)
|
|
|
|
}
|
|
|
|
if h.HealthChecks.Active == nil {
|
|
|
|
h.HealthChecks.Active = new(ActiveHealthChecks)
|
|
|
|
}
|
|
|
|
h.HealthChecks.Active.ExpectBody = d.Val()
|
|
|
|
|
|
|
|
case "max_fails":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if h.HealthChecks == nil {
|
|
|
|
h.HealthChecks = new(HealthChecks)
|
|
|
|
}
|
|
|
|
if h.HealthChecks.Passive == nil {
|
|
|
|
h.HealthChecks.Passive = new(PassiveHealthChecks)
|
|
|
|
}
|
|
|
|
maxFails, err := strconv.Atoi(d.Val())
|
|
|
|
if err != nil {
|
|
|
|
return d.Errf("invalid maximum fail count '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
h.HealthChecks.Passive.MaxFails = maxFails
|
|
|
|
|
|
|
|
case "fail_duration":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if h.HealthChecks == nil {
|
|
|
|
h.HealthChecks = new(HealthChecks)
|
|
|
|
}
|
|
|
|
if h.HealthChecks.Passive == nil {
|
|
|
|
h.HealthChecks.Passive = new(PassiveHealthChecks)
|
|
|
|
}
|
2020-05-12 01:41:11 +03:00
|
|
|
dur, err := caddy.ParseDuration(d.Val())
|
2019-09-09 21:23:27 +03:00
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad duration value '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
h.HealthChecks.Passive.FailDuration = caddy.Duration(dur)
|
|
|
|
|
|
|
|
case "unhealthy_request_count":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if h.HealthChecks == nil {
|
|
|
|
h.HealthChecks = new(HealthChecks)
|
|
|
|
}
|
|
|
|
if h.HealthChecks.Passive == nil {
|
|
|
|
h.HealthChecks.Passive = new(PassiveHealthChecks)
|
|
|
|
}
|
|
|
|
maxConns, err := strconv.Atoi(d.Val())
|
|
|
|
if err != nil {
|
|
|
|
return d.Errf("invalid maximum connection count '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
h.HealthChecks.Passive.UnhealthyRequestCount = maxConns
|
|
|
|
|
|
|
|
case "unhealthy_status":
|
|
|
|
args := d.RemainingArgs()
|
|
|
|
if len(args) == 0 {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if h.HealthChecks == nil {
|
|
|
|
h.HealthChecks = new(HealthChecks)
|
|
|
|
}
|
|
|
|
if h.HealthChecks.Passive == nil {
|
|
|
|
h.HealthChecks.Passive = new(PassiveHealthChecks)
|
|
|
|
}
|
|
|
|
for _, arg := range args {
|
|
|
|
if len(arg) == 3 && strings.HasSuffix(arg, "xx") {
|
|
|
|
arg = arg[:1]
|
|
|
|
}
|
2021-03-30 03:36:40 +03:00
|
|
|
statusNum, err := strconv.Atoi(arg)
|
2019-09-09 21:23:27 +03:00
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad status value '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
h.HealthChecks.Passive.UnhealthyStatus = append(h.HealthChecks.Passive.UnhealthyStatus, statusNum)
|
|
|
|
}
|
|
|
|
|
|
|
|
case "unhealthy_latency":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if h.HealthChecks == nil {
|
|
|
|
h.HealthChecks = new(HealthChecks)
|
|
|
|
}
|
|
|
|
if h.HealthChecks.Passive == nil {
|
|
|
|
h.HealthChecks.Passive = new(PassiveHealthChecks)
|
|
|
|
}
|
2020-05-12 01:41:11 +03:00
|
|
|
dur, err := caddy.ParseDuration(d.Val())
|
2019-09-09 21:23:27 +03:00
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad duration value '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
h.HealthChecks.Passive.UnhealthyLatency = caddy.Duration(dur)
|
|
|
|
|
2019-11-27 21:51:32 +03:00
|
|
|
case "flush_interval":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
2020-01-22 19:34:16 +03:00
|
|
|
if fi, err := strconv.Atoi(d.Val()); err == nil {
|
|
|
|
h.FlushInterval = caddy.Duration(fi)
|
|
|
|
} else {
|
2020-05-12 01:41:11 +03:00
|
|
|
dur, err := caddy.ParseDuration(d.Val())
|
2020-01-22 19:34:16 +03:00
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad duration value '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
h.FlushInterval = caddy.Duration(dur)
|
2019-11-27 21:51:32 +03:00
|
|
|
}
|
|
|
|
|
2020-09-08 19:37:46 +03:00
|
|
|
case "buffer_requests":
|
|
|
|
if d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
h.BufferRequests = true
|
|
|
|
|
2021-02-10 00:15:04 +03:00
|
|
|
case "buffer_responses":
|
|
|
|
if d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
h.BufferResponses = true
|
|
|
|
|
|
|
|
case "max_buffer_size":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
size, err := strconv.Atoi(d.Val())
|
|
|
|
if err != nil {
|
|
|
|
return d.Errf("invalid size (bytes): %s", d.Val())
|
|
|
|
}
|
|
|
|
if d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
h.MaxBufferSize = int64(size)
|
|
|
|
|
2022-03-07 02:51:55 +03:00
|
|
|
case "trusted_proxies":
|
|
|
|
for d.NextArg() {
|
|
|
|
if d.Val() == "private_ranges" {
|
|
|
|
h.TrustedProxies = append(h.TrustedProxies, []string{
|
|
|
|
"192.168.0.0/16",
|
|
|
|
"172.16.0.0/12",
|
|
|
|
"10.0.0.0/8",
|
|
|
|
"127.0.0.1/8",
|
|
|
|
"fd00::/8",
|
|
|
|
"::1",
|
|
|
|
}...)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
h.TrustedProxies = append(h.TrustedProxies, d.Val())
|
|
|
|
}
|
|
|
|
|
2019-09-20 22:13:49 +03:00
|
|
|
case "header_up":
|
2020-11-20 22:38:16 +03:00
|
|
|
var err error
|
|
|
|
|
2019-09-20 22:13:49 +03:00
|
|
|
if h.Headers == nil {
|
|
|
|
h.Headers = new(headers.Handler)
|
|
|
|
}
|
|
|
|
if h.Headers.Request == nil {
|
|
|
|
h.Headers.Request = new(headers.HeaderOps)
|
|
|
|
}
|
|
|
|
args := d.RemainingArgs()
|
2020-11-20 22:38:16 +03:00
|
|
|
|
2019-09-20 22:13:49 +03:00
|
|
|
switch len(args) {
|
|
|
|
case 1:
|
2020-11-20 22:38:16 +03:00
|
|
|
err = headers.CaddyfileHeaderOp(h.Headers.Request, args[0], "", "")
|
2019-09-20 22:13:49 +03:00
|
|
|
case 2:
|
2021-01-04 21:11:36 +03:00
|
|
|
// some lint checks, I guess
|
|
|
|
if strings.EqualFold(args[0], "host") && (args[1] == "{hostport}" || args[1] == "{http.request.hostport}") {
|
|
|
|
log.Printf("[WARNING] Unnecessary header_up ('Host' field): the reverse proxy's default behavior is to pass headers to the upstream")
|
|
|
|
}
|
2022-03-07 02:51:55 +03:00
|
|
|
if strings.EqualFold(args[0], "x-forwarded-for") && (args[1] == "{remote}" || args[1] == "{http.request.remote}" || args[1] == "{remote_host}" || args[1] == "{http.request.remote.host}") {
|
|
|
|
log.Printf("[WARNING] Unnecessary header_up ('X-Forwarded-For' field): the reverse proxy's default behavior is to pass headers to the upstream")
|
|
|
|
}
|
2021-01-04 21:11:36 +03:00
|
|
|
if strings.EqualFold(args[0], "x-forwarded-proto") && (args[1] == "{scheme}" || args[1] == "{http.request.scheme}") {
|
|
|
|
log.Printf("[WARNING] Unnecessary header_up ('X-Forwarded-Proto' field): the reverse proxy's default behavior is to pass headers to the upstream")
|
|
|
|
}
|
2022-03-07 02:51:55 +03:00
|
|
|
if strings.EqualFold(args[0], "x-forwarded-host") && (args[1] == "{host}" || args[1] == "{http.request.host}" || args[1] == "{hostport}" || args[1] == "{http.request.hostport}") {
|
|
|
|
log.Printf("[WARNING] Unnecessary header_up ('X-Forwarded-Host' field): the reverse proxy's default behavior is to pass headers to the upstream")
|
|
|
|
}
|
2020-11-20 22:38:16 +03:00
|
|
|
err = headers.CaddyfileHeaderOp(h.Headers.Request, args[0], args[1], "")
|
2019-09-20 22:13:49 +03:00
|
|
|
case 3:
|
2020-11-20 22:38:16 +03:00
|
|
|
err = headers.CaddyfileHeaderOp(h.Headers.Request, args[0], args[1], args[2])
|
2019-09-20 22:13:49 +03:00
|
|
|
default:
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
|
2020-11-20 22:38:16 +03:00
|
|
|
if err != nil {
|
|
|
|
return d.Err(err.Error())
|
|
|
|
}
|
|
|
|
|
2019-09-20 22:13:49 +03:00
|
|
|
case "header_down":
|
2020-11-20 22:38:16 +03:00
|
|
|
var err error
|
|
|
|
|
2019-09-20 22:13:49 +03:00
|
|
|
if h.Headers == nil {
|
|
|
|
h.Headers = new(headers.Handler)
|
|
|
|
}
|
|
|
|
if h.Headers.Response == nil {
|
|
|
|
h.Headers.Response = &headers.RespHeaderOps{
|
|
|
|
HeaderOps: new(headers.HeaderOps),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
args := d.RemainingArgs()
|
|
|
|
switch len(args) {
|
|
|
|
case 1:
|
2020-11-20 22:38:16 +03:00
|
|
|
err = headers.CaddyfileHeaderOp(h.Headers.Response.HeaderOps, args[0], "", "")
|
2019-09-20 22:13:49 +03:00
|
|
|
case 2:
|
2020-11-20 22:38:16 +03:00
|
|
|
err = headers.CaddyfileHeaderOp(h.Headers.Response.HeaderOps, args[0], args[1], "")
|
2019-09-20 22:13:49 +03:00
|
|
|
case 3:
|
2020-11-20 22:38:16 +03:00
|
|
|
err = headers.CaddyfileHeaderOp(h.Headers.Response.HeaderOps, args[0], args[1], args[2])
|
2019-09-20 22:13:49 +03:00
|
|
|
default:
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
|
2020-11-20 22:38:16 +03:00
|
|
|
if err != nil {
|
|
|
|
return d.Err(err.Error())
|
|
|
|
}
|
|
|
|
|
2019-09-09 21:23:27 +03:00
|
|
|
case "transport":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if h.TransportRaw != nil {
|
|
|
|
return d.Err("transport already specified")
|
|
|
|
}
|
2020-02-28 06:56:24 +03:00
|
|
|
transportModuleName = d.Val()
|
2021-01-06 00:39:30 +03:00
|
|
|
modID := "http.reverse_proxy.transport." + transportModuleName
|
|
|
|
unm, err := caddyfile.UnmarshalModule(d, modID)
|
2019-09-09 21:23:27 +03:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
rt, ok := unm.(http.RoundTripper)
|
|
|
|
if !ok {
|
2021-01-06 00:39:30 +03:00
|
|
|
return d.Errf("module %s (%T) is not a RoundTripper", modID, unm)
|
2019-09-09 21:23:27 +03:00
|
|
|
}
|
2020-02-28 06:56:24 +03:00
|
|
|
transport = rt
|
2019-09-09 21:23:27 +03:00
|
|
|
|
2021-05-02 21:39:06 +03:00
|
|
|
case "handle_response":
|
|
|
|
// delegate the parsing of handle_response to the caller,
|
|
|
|
// since we need the httpcaddyfile.Helper to parse subroutes.
|
|
|
|
// See h.FinalizeUnmarshalCaddyfile
|
|
|
|
h.handleResponseSegments = append(h.handleResponseSegments, d.NewFromNextSegment())
|
|
|
|
|
2022-03-02 00:12:43 +03:00
|
|
|
case "replace_status":
|
|
|
|
args := d.RemainingArgs()
|
|
|
|
if len(args) != 2 {
|
|
|
|
return d.Errf("must have two arguments: a response matcher and a status code")
|
|
|
|
}
|
|
|
|
|
|
|
|
if !strings.HasPrefix(args[0], matcherPrefix) {
|
|
|
|
return d.Errf("must use a named response matcher, starting with '@'")
|
|
|
|
}
|
|
|
|
|
|
|
|
foundMatcher, ok := h.responseMatchers[args[0]]
|
|
|
|
if !ok {
|
|
|
|
return d.Errf("no named response matcher defined with name '%s'", args[0][1:])
|
|
|
|
}
|
|
|
|
|
|
|
|
_, err := strconv.Atoi(args[1])
|
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad integer value '%s': %v", args[1], err)
|
|
|
|
}
|
|
|
|
|
|
|
|
// make sure there's no block, cause it doesn't make sense
|
|
|
|
if d.NextBlock(1) {
|
|
|
|
return d.Errf("cannot define routes for 'replace_status', use 'handle_response' instead.")
|
|
|
|
}
|
|
|
|
|
|
|
|
h.HandleResponse = append(
|
|
|
|
h.HandleResponse,
|
|
|
|
caddyhttp.ResponseHandler{
|
|
|
|
Match: &foundMatcher,
|
|
|
|
StatusCode: caddyhttp.WeakString(args[1]),
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
2019-09-09 21:23:27 +03:00
|
|
|
default:
|
|
|
|
return d.Errf("unrecognized subdirective %s", d.Val())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-02-28 06:56:24 +03:00
|
|
|
// if the scheme inferred from the backends' addresses is
|
2020-11-23 22:18:26 +03:00
|
|
|
// HTTPS, we will need a non-nil transport to enable TLS,
|
|
|
|
// or if H2C, to set the transport versions.
|
|
|
|
if (commonScheme == "https" || commonScheme == "h2c") && transport == nil {
|
2020-02-28 06:56:24 +03:00
|
|
|
transport = new(HTTPTransport)
|
|
|
|
transportModuleName = "http"
|
|
|
|
}
|
|
|
|
|
|
|
|
// verify transport configuration, and finally encode it
|
|
|
|
if transport != nil {
|
2020-04-07 17:31:52 +03:00
|
|
|
if te, ok := transport.(TLSTransport); ok {
|
|
|
|
if commonScheme == "https" && !te.TLSEnabled() {
|
|
|
|
err := te.EnableTLS(new(TLSConfig))
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2020-02-28 06:56:24 +03:00
|
|
|
}
|
2020-04-07 17:31:52 +03:00
|
|
|
if commonScheme == "http" && te.TLSEnabled() {
|
2020-02-28 06:56:24 +03:00
|
|
|
return d.Errf("upstream address scheme is HTTP but transport is configured for HTTP+TLS (HTTPS)")
|
|
|
|
}
|
2020-11-23 22:18:26 +03:00
|
|
|
if te, ok := transport.(*HTTPTransport); ok && commonScheme == "h2c" {
|
|
|
|
te.Versions = []string{"h2c", "2"}
|
|
|
|
}
|
2020-04-07 17:31:52 +03:00
|
|
|
} else if commonScheme == "https" {
|
|
|
|
return d.Errf("upstreams are configured for HTTPS but transport module does not support TLS: %T", transport)
|
2020-02-28 06:56:24 +03:00
|
|
|
}
|
2020-07-17 22:18:32 +03:00
|
|
|
|
|
|
|
// no need to encode empty default transport
|
|
|
|
if !reflect.DeepEqual(transport, new(HTTPTransport)) {
|
2020-02-28 06:56:24 +03:00
|
|
|
h.TransportRaw = caddyconfig.JSONModuleObject(transport, "protocol", transportModuleName, nil)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-09-09 21:23:27 +03:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2021-05-02 21:39:06 +03:00
|
|
|
// FinalizeUnmarshalCaddyfile finalizes the Caddyfile parsing which
|
|
|
|
// requires having an httpcaddyfile.Helper to function, to parse subroutes.
|
|
|
|
func (h *Handler) FinalizeUnmarshalCaddyfile(helper httpcaddyfile.Helper) error {
|
|
|
|
for _, d := range h.handleResponseSegments {
|
|
|
|
// consume the "handle_response" token
|
|
|
|
d.Next()
|
|
|
|
args := d.RemainingArgs()
|
|
|
|
|
2022-03-02 00:12:43 +03:00
|
|
|
// TODO: Remove this check at some point in the future
|
|
|
|
if len(args) == 2 {
|
|
|
|
return d.Errf("configuring 'handle_response' for status code replacement is no longer supported. Use 'replace_status' instead.")
|
|
|
|
}
|
|
|
|
|
|
|
|
if len(args) > 1 {
|
2021-05-02 21:39:06 +03:00
|
|
|
return d.Errf("too many arguments for 'handle_response': %s", args)
|
|
|
|
}
|
|
|
|
|
2022-03-02 00:12:43 +03:00
|
|
|
var matcher *caddyhttp.ResponseMatcher
|
|
|
|
if len(args) == 1 {
|
|
|
|
// the first arg should always be a matcher.
|
2021-05-02 21:39:06 +03:00
|
|
|
if !strings.HasPrefix(args[0], matcherPrefix) {
|
|
|
|
return d.Errf("must use a named response matcher, starting with '@'")
|
|
|
|
}
|
|
|
|
|
|
|
|
foundMatcher, ok := h.responseMatchers[args[0]]
|
|
|
|
if !ok {
|
|
|
|
return d.Errf("no named response matcher defined with name '%s'", args[0][1:])
|
|
|
|
}
|
|
|
|
matcher = &foundMatcher
|
|
|
|
}
|
|
|
|
|
|
|
|
// parse the block as routes
|
|
|
|
handler, err := httpcaddyfile.ParseSegmentAsSubroute(helper.WithDispenser(d.NewFromNextSegment()))
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
subroute, ok := handler.(*caddyhttp.Subroute)
|
|
|
|
if !ok {
|
|
|
|
return helper.Errf("segment was not parsed as a subroute")
|
|
|
|
}
|
|
|
|
h.HandleResponse = append(
|
|
|
|
h.HandleResponse,
|
|
|
|
caddyhttp.ResponseHandler{
|
|
|
|
Match: matcher,
|
|
|
|
Routes: subroute.Routes,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
}
|
|
|
|
|
|
|
|
// move the handle_response entries without a matcher to the end.
|
|
|
|
// we can't use sort.SliceStable because it will reorder the rest of the
|
|
|
|
// entries which may be undesirable because we don't have a good
|
|
|
|
// heuristic to use for sorting.
|
|
|
|
withoutMatchers := []caddyhttp.ResponseHandler{}
|
|
|
|
withMatchers := []caddyhttp.ResponseHandler{}
|
|
|
|
for _, hr := range h.HandleResponse {
|
|
|
|
if hr.Match == nil {
|
|
|
|
withoutMatchers = append(withoutMatchers, hr)
|
|
|
|
} else {
|
|
|
|
withMatchers = append(withMatchers, hr)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
h.HandleResponse = append(withMatchers, withoutMatchers...)
|
|
|
|
|
|
|
|
// clean up the bits we only needed for adapting
|
|
|
|
h.handleResponseSegments = nil
|
|
|
|
h.responseMatchers = nil
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2019-09-09 21:23:27 +03:00
|
|
|
// UnmarshalCaddyfile deserializes Caddyfile tokens into h.
|
|
|
|
//
|
|
|
|
// transport http {
|
2020-11-03 00:59:02 +03:00
|
|
|
// read_buffer <size>
|
|
|
|
// write_buffer <size>
|
|
|
|
// max_response_header <size>
|
|
|
|
// dial_timeout <duration>
|
|
|
|
// dial_fallback_delay <duration>
|
|
|
|
// response_header_timeout <duration>
|
|
|
|
// expect_continue_timeout <duration>
|
2022-03-07 03:43:39 +03:00
|
|
|
// resolvers <resolvers...>
|
2020-10-30 21:05:21 +03:00
|
|
|
// tls
|
|
|
|
// tls_client_auth <automate_name> | <cert_file> <key_file>
|
2019-09-09 21:23:27 +03:00
|
|
|
// tls_insecure_skip_verify
|
|
|
|
// tls_timeout <duration>
|
2020-01-07 22:07:42 +03:00
|
|
|
// tls_trusted_ca_certs <cert_files...>
|
2020-10-30 21:05:21 +03:00
|
|
|
// tls_server_name <sni>
|
2019-09-09 21:23:27 +03:00
|
|
|
// keepalive [off|<duration>]
|
2021-11-24 09:32:25 +03:00
|
|
|
// keepalive_interval <interval>
|
2019-09-09 21:23:27 +03:00
|
|
|
// keepalive_idle_conns <max_count>
|
2021-11-24 09:32:25 +03:00
|
|
|
// keepalive_idle_conns_per_host <count>
|
2020-05-05 21:33:21 +03:00
|
|
|
// versions <versions...>
|
2020-10-30 21:05:21 +03:00
|
|
|
// compression off
|
|
|
|
// max_conns_per_host <count>
|
|
|
|
// max_idle_conns_per_host <count>
|
2019-09-09 21:23:27 +03:00
|
|
|
// }
|
|
|
|
//
|
|
|
|
func (h *HTTPTransport) UnmarshalCaddyfile(d *caddyfile.Dispenser) error {
|
2019-09-12 03:53:44 +03:00
|
|
|
for d.Next() {
|
|
|
|
for d.NextBlock(0) {
|
|
|
|
switch d.Val() {
|
|
|
|
case "read_buffer":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
size, err := humanize.ParseBytes(d.Val())
|
|
|
|
if err != nil {
|
|
|
|
return d.Errf("invalid read buffer size '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
h.ReadBufferSize = int(size)
|
2019-09-09 21:23:27 +03:00
|
|
|
|
2019-09-12 03:53:44 +03:00
|
|
|
case "write_buffer":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
size, err := humanize.ParseBytes(d.Val())
|
|
|
|
if err != nil {
|
|
|
|
return d.Errf("invalid write buffer size '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
h.WriteBufferSize = int(size)
|
2019-09-09 21:23:27 +03:00
|
|
|
|
2020-11-03 00:59:02 +03:00
|
|
|
case "max_response_header":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
size, err := humanize.ParseBytes(d.Val())
|
|
|
|
if err != nil {
|
|
|
|
return d.Errf("invalid max response header size '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
h.MaxResponseHeaderSize = int64(size)
|
|
|
|
|
2019-09-12 03:53:44 +03:00
|
|
|
case "dial_timeout":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
2020-05-12 01:41:11 +03:00
|
|
|
dur, err := caddy.ParseDuration(d.Val())
|
2019-09-12 03:53:44 +03:00
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad timeout value '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
h.DialTimeout = caddy.Duration(dur)
|
2019-09-09 21:23:27 +03:00
|
|
|
|
2020-11-03 00:59:02 +03:00
|
|
|
case "dial_fallback_delay":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
dur, err := caddy.ParseDuration(d.Val())
|
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad fallback delay value '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
h.FallbackDelay = caddy.Duration(dur)
|
|
|
|
|
|
|
|
case "response_header_timeout":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
dur, err := caddy.ParseDuration(d.Val())
|
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad timeout value '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
h.ResponseHeaderTimeout = caddy.Duration(dur)
|
|
|
|
|
|
|
|
case "expect_continue_timeout":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
dur, err := caddy.ParseDuration(d.Val())
|
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad timeout value '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
h.ExpectContinueTimeout = caddy.Duration(dur)
|
|
|
|
|
2022-03-07 03:43:39 +03:00
|
|
|
case "resolvers":
|
|
|
|
if h.Resolver == nil {
|
|
|
|
h.Resolver = new(UpstreamResolver)
|
|
|
|
}
|
|
|
|
h.Resolver.Addresses = d.RemainingArgs()
|
|
|
|
if len(h.Resolver.Addresses) == 0 {
|
|
|
|
return d.Errf("must specify at least one resolver address")
|
|
|
|
}
|
|
|
|
|
2019-09-12 03:53:44 +03:00
|
|
|
case "tls_client_auth":
|
|
|
|
if h.TLS == nil {
|
|
|
|
h.TLS = new(TLSConfig)
|
|
|
|
}
|
2020-06-08 19:30:26 +03:00
|
|
|
args := d.RemainingArgs()
|
|
|
|
switch len(args) {
|
|
|
|
case 1:
|
|
|
|
h.TLS.ClientCertificateAutomate = args[0]
|
|
|
|
case 2:
|
|
|
|
h.TLS.ClientCertificateFile = args[0]
|
|
|
|
h.TLS.ClientCertificateKeyFile = args[1]
|
|
|
|
default:
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
2019-09-09 21:23:27 +03:00
|
|
|
|
2019-09-12 03:53:44 +03:00
|
|
|
case "tls":
|
|
|
|
if h.TLS == nil {
|
|
|
|
h.TLS = new(TLSConfig)
|
|
|
|
}
|
2019-09-12 03:46:32 +03:00
|
|
|
|
2019-09-12 03:53:44 +03:00
|
|
|
case "tls_insecure_skip_verify":
|
|
|
|
if d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if h.TLS == nil {
|
|
|
|
h.TLS = new(TLSConfig)
|
|
|
|
}
|
|
|
|
h.TLS.InsecureSkipVerify = true
|
2019-09-09 21:23:27 +03:00
|
|
|
|
2019-09-12 03:53:44 +03:00
|
|
|
case "tls_timeout":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
2020-05-12 01:41:11 +03:00
|
|
|
dur, err := caddy.ParseDuration(d.Val())
|
2019-09-12 03:53:44 +03:00
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad timeout value '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
if h.TLS == nil {
|
|
|
|
h.TLS = new(TLSConfig)
|
|
|
|
}
|
|
|
|
h.TLS.HandshakeTimeout = caddy.Duration(dur)
|
2019-09-09 21:23:27 +03:00
|
|
|
|
2020-01-07 22:07:42 +03:00
|
|
|
case "tls_trusted_ca_certs":
|
|
|
|
args := d.RemainingArgs()
|
|
|
|
if len(args) == 0 {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if h.TLS == nil {
|
|
|
|
h.TLS = new(TLSConfig)
|
|
|
|
}
|
2020-01-22 19:34:16 +03:00
|
|
|
h.TLS.RootCAPEMFiles = args
|
2020-01-07 22:07:42 +03:00
|
|
|
|
2020-05-05 21:39:39 +03:00
|
|
|
case "tls_server_name":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if h.TLS == nil {
|
|
|
|
h.TLS = new(TLSConfig)
|
|
|
|
}
|
|
|
|
h.TLS.ServerName = d.Val()
|
|
|
|
|
2019-09-12 03:53:44 +03:00
|
|
|
case "keepalive":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if h.KeepAlive == nil {
|
|
|
|
h.KeepAlive = new(KeepAlive)
|
|
|
|
}
|
|
|
|
if d.Val() == "off" {
|
|
|
|
var disable bool
|
|
|
|
h.KeepAlive.Enabled = &disable
|
2019-10-11 23:25:39 +03:00
|
|
|
break
|
2019-09-12 03:53:44 +03:00
|
|
|
}
|
2020-05-12 01:41:11 +03:00
|
|
|
dur, err := caddy.ParseDuration(d.Val())
|
2019-09-12 03:53:44 +03:00
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad duration value '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
h.KeepAlive.IdleConnTimeout = caddy.Duration(dur)
|
2019-09-09 21:23:27 +03:00
|
|
|
|
2021-11-24 09:32:25 +03:00
|
|
|
case "keepalive_interval":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
dur, err := caddy.ParseDuration(d.Val())
|
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad interval value '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
if h.KeepAlive == nil {
|
|
|
|
h.KeepAlive = new(KeepAlive)
|
|
|
|
}
|
|
|
|
h.KeepAlive.ProbeInterval = caddy.Duration(dur)
|
|
|
|
|
2019-09-12 03:53:44 +03:00
|
|
|
case "keepalive_idle_conns":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
num, err := strconv.Atoi(d.Val())
|
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad integer value '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
if h.KeepAlive == nil {
|
|
|
|
h.KeepAlive = new(KeepAlive)
|
|
|
|
}
|
|
|
|
h.KeepAlive.MaxIdleConns = num
|
2021-06-15 23:54:48 +03:00
|
|
|
|
|
|
|
case "keepalive_idle_conns_per_host":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
num, err := strconv.Atoi(d.Val())
|
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad integer value '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
if h.KeepAlive == nil {
|
|
|
|
h.KeepAlive = new(KeepAlive)
|
|
|
|
}
|
2019-10-11 23:25:39 +03:00
|
|
|
h.KeepAlive.MaxIdleConnsPerHost = num
|
2019-09-09 21:23:27 +03:00
|
|
|
|
2020-05-05 21:33:21 +03:00
|
|
|
case "versions":
|
|
|
|
h.Versions = d.RemainingArgs()
|
|
|
|
if len(h.Versions) == 0 {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
|
2020-07-31 20:30:20 +03:00
|
|
|
case "compression":
|
|
|
|
if d.NextArg() {
|
|
|
|
if d.Val() == "off" {
|
|
|
|
var disable bool
|
|
|
|
h.Compression = &disable
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-10-30 21:05:21 +03:00
|
|
|
case "max_conns_per_host":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
num, err := strconv.Atoi(d.Val())
|
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad integer value '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
h.MaxConnsPerHost = num
|
|
|
|
|
2019-09-12 03:53:44 +03:00
|
|
|
default:
|
|
|
|
return d.Errf("unrecognized subdirective %s", d.Val())
|
|
|
|
}
|
2019-09-09 21:23:27 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
reverseproxy: copy_response and copy_response_headers for handle_response routes (#4391)
* reverseproxy: New `copy_response` handler for `handle_response` routes
Followup to #4298 and #4388.
This adds a new `copy_response` handler which may only be used in `reverse_proxy`'s `handle_response` routes, which can be used to actually copy the proxy response downstream.
Previously, if `handle_response` was used (with routes, not the status code mode), it was impossible to use the upstream's response body at all, because we would always close the body, expecting the routes to write a new body from scratch.
To implement this, I had to refactor `h.reverseProxy()` to move all the code that came after the `HandleResponse` loop into a new function. This new function `h.finalizeResponse()` takes care of preparing the response by removing extra headers, dealing with trailers, then copying the headers and body downstream.
Since basically what we want `copy_response` to do is invoke `h.finalizeResponse()` at a configurable point in time, we need to pass down the proxy handler, the response, and some other state via a new `req.WithContext(ctx)`. Wrapping a new context is pretty much the only way we have to jump a few layers in the HTTP middleware chain and let a handler pick up this information. Feels a bit dirty, but it works.
Also fixed a bug with the `http.reverse_proxy.upstream.duration` placeholder, it always had the same duration as `http.reverse_proxy.upstream.latency`, but the former was meant to be the time taken for the roundtrip _plus_ copying/writing the response.
* Delete the "Content-Length" header if we aren't copying
Fixes a bug where the Content-Length will mismatch the actual bytes written if we skipped copying the response, so we get a message like this when using curl:
```
curl: (18) transfer closed with 18 bytes remaining to read
```
To replicate:
```
{
admin off
debug
}
:8881 {
reverse_proxy 127.0.0.1:8882 {
@200 status 200
handle_response @200 {
header Foo bar
}
}
}
:8882 {
header Content-Type application/json
respond `{"hello": "world"}` 200
}
```
* Implement `copy_response_headers`, with include/exclude list support
* Apply suggestions from code review
Co-authored-by: Matt Holt <mholt@users.noreply.github.com>
2022-03-09 21:00:51 +03:00
|
|
|
func parseCopyResponseCaddyfile(h httpcaddyfile.Helper) (caddyhttp.MiddlewareHandler, error) {
|
|
|
|
crh := new(CopyResponseHandler)
|
|
|
|
err := crh.UnmarshalCaddyfile(h.Dispenser)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
return crh, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// UnmarshalCaddyfile sets up the handler from Caddyfile tokens. Syntax:
|
|
|
|
//
|
|
|
|
// copy_response [<matcher>] [<status>] {
|
|
|
|
// status <status>
|
|
|
|
// }
|
|
|
|
//
|
|
|
|
func (h *CopyResponseHandler) UnmarshalCaddyfile(d *caddyfile.Dispenser) error {
|
|
|
|
for d.Next() {
|
|
|
|
args := d.RemainingArgs()
|
|
|
|
if len(args) == 1 {
|
|
|
|
if num, err := strconv.Atoi(args[0]); err == nil && num > 0 {
|
|
|
|
h.StatusCode = caddyhttp.WeakString(args[0])
|
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
for d.NextBlock(0) {
|
|
|
|
switch d.Val() {
|
|
|
|
case "status":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
h.StatusCode = caddyhttp.WeakString(d.Val())
|
|
|
|
default:
|
|
|
|
return d.Errf("unrecognized subdirective '%s'", d.Val())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func parseCopyResponseHeadersCaddyfile(h httpcaddyfile.Helper) (caddyhttp.MiddlewareHandler, error) {
|
|
|
|
crh := new(CopyResponseHeadersHandler)
|
|
|
|
err := crh.UnmarshalCaddyfile(h.Dispenser)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
return crh, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// UnmarshalCaddyfile sets up the handler from Caddyfile tokens. Syntax:
|
|
|
|
//
|
|
|
|
// copy_response_headers [<matcher>] {
|
2022-03-13 09:38:11 +03:00
|
|
|
// include <fields...>
|
reverseproxy: copy_response and copy_response_headers for handle_response routes (#4391)
* reverseproxy: New `copy_response` handler for `handle_response` routes
Followup to #4298 and #4388.
This adds a new `copy_response` handler which may only be used in `reverse_proxy`'s `handle_response` routes, which can be used to actually copy the proxy response downstream.
Previously, if `handle_response` was used (with routes, not the status code mode), it was impossible to use the upstream's response body at all, because we would always close the body, expecting the routes to write a new body from scratch.
To implement this, I had to refactor `h.reverseProxy()` to move all the code that came after the `HandleResponse` loop into a new function. This new function `h.finalizeResponse()` takes care of preparing the response by removing extra headers, dealing with trailers, then copying the headers and body downstream.
Since basically what we want `copy_response` to do is invoke `h.finalizeResponse()` at a configurable point in time, we need to pass down the proxy handler, the response, and some other state via a new `req.WithContext(ctx)`. Wrapping a new context is pretty much the only way we have to jump a few layers in the HTTP middleware chain and let a handler pick up this information. Feels a bit dirty, but it works.
Also fixed a bug with the `http.reverse_proxy.upstream.duration` placeholder, it always had the same duration as `http.reverse_proxy.upstream.latency`, but the former was meant to be the time taken for the roundtrip _plus_ copying/writing the response.
* Delete the "Content-Length" header if we aren't copying
Fixes a bug where the Content-Length will mismatch the actual bytes written if we skipped copying the response, so we get a message like this when using curl:
```
curl: (18) transfer closed with 18 bytes remaining to read
```
To replicate:
```
{
admin off
debug
}
:8881 {
reverse_proxy 127.0.0.1:8882 {
@200 status 200
handle_response @200 {
header Foo bar
}
}
}
:8882 {
header Content-Type application/json
respond `{"hello": "world"}` 200
}
```
* Implement `copy_response_headers`, with include/exclude list support
* Apply suggestions from code review
Co-authored-by: Matt Holt <mholt@users.noreply.github.com>
2022-03-09 21:00:51 +03:00
|
|
|
// exclude <fields...>
|
|
|
|
// }
|
|
|
|
//
|
|
|
|
func (h *CopyResponseHeadersHandler) UnmarshalCaddyfile(d *caddyfile.Dispenser) error {
|
|
|
|
for d.Next() {
|
|
|
|
args := d.RemainingArgs()
|
|
|
|
if len(args) > 0 {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
|
|
|
|
for d.NextBlock(0) {
|
|
|
|
switch d.Val() {
|
|
|
|
case "include":
|
|
|
|
h.Include = append(h.Include, d.RemainingArgs()...)
|
|
|
|
|
|
|
|
case "exclude":
|
|
|
|
h.Exclude = append(h.Exclude, d.RemainingArgs()...)
|
|
|
|
|
|
|
|
default:
|
|
|
|
return d.Errf("unrecognized subdirective '%s'", d.Val())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2022-03-07 03:43:39 +03:00
|
|
|
// UnmarshalCaddyfile deserializes Caddyfile tokens into h.
|
|
|
|
//
|
|
|
|
// dynamic srv [<name>] {
|
|
|
|
// service <service>
|
|
|
|
// proto <proto>
|
|
|
|
// name <name>
|
|
|
|
// refresh <interval>
|
|
|
|
// resolvers <resolvers...>
|
|
|
|
// dial_timeout <timeout>
|
|
|
|
// dial_fallback_delay <timeout>
|
|
|
|
// }
|
|
|
|
//
|
|
|
|
func (u *SRVUpstreams) UnmarshalCaddyfile(d *caddyfile.Dispenser) error {
|
|
|
|
for d.Next() {
|
|
|
|
args := d.RemainingArgs()
|
|
|
|
if len(args) > 1 {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if len(args) > 0 {
|
|
|
|
u.Name = args[0]
|
|
|
|
}
|
|
|
|
|
|
|
|
for d.NextBlock(0) {
|
|
|
|
switch d.Val() {
|
|
|
|
case "service":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if u.Service != "" {
|
|
|
|
return d.Errf("srv service has already been specified")
|
|
|
|
}
|
|
|
|
u.Service = d.Val()
|
|
|
|
|
|
|
|
case "proto":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if u.Proto != "" {
|
|
|
|
return d.Errf("srv proto has already been specified")
|
|
|
|
}
|
|
|
|
u.Proto = d.Val()
|
|
|
|
|
|
|
|
case "name":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if u.Name != "" {
|
|
|
|
return d.Errf("srv name has already been specified")
|
|
|
|
}
|
|
|
|
u.Name = d.Val()
|
|
|
|
|
|
|
|
case "refresh":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
dur, err := caddy.ParseDuration(d.Val())
|
|
|
|
if err != nil {
|
|
|
|
return d.Errf("parsing refresh interval duration: %v", err)
|
|
|
|
}
|
|
|
|
u.Refresh = caddy.Duration(dur)
|
|
|
|
|
|
|
|
case "resolvers":
|
|
|
|
if u.Resolver == nil {
|
|
|
|
u.Resolver = new(UpstreamResolver)
|
|
|
|
}
|
|
|
|
u.Resolver.Addresses = d.RemainingArgs()
|
|
|
|
if len(u.Resolver.Addresses) == 0 {
|
|
|
|
return d.Errf("must specify at least one resolver address")
|
|
|
|
}
|
|
|
|
|
|
|
|
case "dial_timeout":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
dur, err := caddy.ParseDuration(d.Val())
|
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad timeout value '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
u.DialTimeout = caddy.Duration(dur)
|
|
|
|
|
|
|
|
case "dial_fallback_delay":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
dur, err := caddy.ParseDuration(d.Val())
|
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad delay value '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
u.FallbackDelay = caddy.Duration(dur)
|
|
|
|
|
|
|
|
default:
|
|
|
|
return d.Errf("unrecognized srv option '%s'", d.Val())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// UnmarshalCaddyfile deserializes Caddyfile tokens into h.
|
|
|
|
//
|
|
|
|
// dynamic a [<name> <port] {
|
|
|
|
// name <name>
|
|
|
|
// port <port>
|
|
|
|
// refresh <interval>
|
|
|
|
// resolvers <resolvers...>
|
|
|
|
// dial_timeout <timeout>
|
|
|
|
// dial_fallback_delay <timeout>
|
|
|
|
// }
|
|
|
|
//
|
|
|
|
func (u *AUpstreams) UnmarshalCaddyfile(d *caddyfile.Dispenser) error {
|
|
|
|
for d.Next() {
|
|
|
|
args := d.RemainingArgs()
|
|
|
|
if len(args) > 2 {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if len(args) > 0 {
|
|
|
|
u.Name = args[0]
|
|
|
|
u.Port = args[1]
|
|
|
|
}
|
|
|
|
|
|
|
|
for d.NextBlock(0) {
|
|
|
|
switch d.Val() {
|
|
|
|
case "name":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if u.Name != "" {
|
|
|
|
return d.Errf("a name has already been specified")
|
|
|
|
}
|
|
|
|
u.Name = d.Val()
|
|
|
|
|
|
|
|
case "port":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
if u.Port != "" {
|
|
|
|
return d.Errf("a port has already been specified")
|
|
|
|
}
|
|
|
|
u.Port = d.Val()
|
|
|
|
|
|
|
|
case "refresh":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
dur, err := caddy.ParseDuration(d.Val())
|
|
|
|
if err != nil {
|
|
|
|
return d.Errf("parsing refresh interval duration: %v", err)
|
|
|
|
}
|
|
|
|
u.Refresh = caddy.Duration(dur)
|
|
|
|
|
|
|
|
case "resolvers":
|
|
|
|
if u.Resolver == nil {
|
|
|
|
u.Resolver = new(UpstreamResolver)
|
|
|
|
}
|
|
|
|
u.Resolver.Addresses = d.RemainingArgs()
|
|
|
|
if len(u.Resolver.Addresses) == 0 {
|
|
|
|
return d.Errf("must specify at least one resolver address")
|
|
|
|
}
|
|
|
|
|
|
|
|
case "dial_timeout":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
dur, err := caddy.ParseDuration(d.Val())
|
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad timeout value '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
u.DialTimeout = caddy.Duration(dur)
|
|
|
|
|
|
|
|
case "dial_fallback_delay":
|
|
|
|
if !d.NextArg() {
|
|
|
|
return d.ArgErr()
|
|
|
|
}
|
|
|
|
dur, err := caddy.ParseDuration(d.Val())
|
|
|
|
if err != nil {
|
|
|
|
return d.Errf("bad delay value '%s': %v", d.Val(), err)
|
|
|
|
}
|
|
|
|
u.FallbackDelay = caddy.Duration(dur)
|
|
|
|
|
|
|
|
default:
|
|
|
|
return d.Errf("unrecognized srv option '%s'", d.Val())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2021-05-02 21:39:06 +03:00
|
|
|
const matcherPrefix = "@"
|
|
|
|
|
2019-09-09 21:23:27 +03:00
|
|
|
// Interface guards
|
|
|
|
var (
|
|
|
|
_ caddyfile.Unmarshaler = (*Handler)(nil)
|
|
|
|
_ caddyfile.Unmarshaler = (*HTTPTransport)(nil)
|
2022-03-07 03:43:39 +03:00
|
|
|
_ caddyfile.Unmarshaler = (*SRVUpstreams)(nil)
|
|
|
|
_ caddyfile.Unmarshaler = (*AUpstreams)(nil)
|
2019-09-09 21:23:27 +03:00
|
|
|
)
|