Commit e9c2e506 authored by Matthew Holt's avatar Matthew Holt

Merge branch 'master' into letsencryptfix

# Conflicts:
#	caddy/letsencrypt/letsencrypt.go
parents 55601d3e b6326d40
...@@ -2,7 +2,7 @@ language: go ...@@ -2,7 +2,7 @@ language: go
go: go:
- 1.4.3 - 1.4.3
- 1.5.1 - 1.5.2
- tip - tip
install: install:
......
## Contributing to Caddy ## Contributing to Caddy
**[Join us on Slack](https://gophers.slack.com/messages/caddy/)** to chat with **[Join our dev chat on Gitter](https://gitter.im/mholt/caddy)** to chat with
other Caddy developers! ([Request an invite](http://bit.ly/go-slack-signup), other Caddy developers! (Dev chat only; try our
then join the #caddy channel.) [support room](https://gitter.im/caddyserver/support) for help or
[general](https://gitter.im/caddyserver/general) for anything else.)
This project gladly accepts contributions and we encourage interested users to This project gladly accepts contributions and we encourage interested users to
get involved! get involved!
...@@ -11,24 +12,25 @@ get involved! ...@@ -11,24 +12,25 @@ get involved!
#### For small tweaks, bug fixes, and tests #### For small tweaks, bug fixes, and tests
Submit [pull requests](https://github.com/mholt/caddy/pulls) at any time. Submit [pull requests](https://github.com/mholt/caddy/pulls) at any time.
Thank you for helping out in simple ways! Bug fixes should be under test to Bug fixes should be under test to assert correct behavior. Thank you for
assert correct behavior. helping out in simple ways!
#### Ideas, questions, bug reports #### Ideas, questions, bug reports
You should totally [open an issue](https://github.com/mholt/caddy/issues) with Feel free to [open an issue](https://github.com/mholt/caddy/issues) with your
your ideas, questions, and bug reports, if one does not already exist for it. ideas, questions, and bug reports, if one does not already exist for it. Bug
Bug reports should state expected behavior and contain clear instructions for reports should state expected behavior and contain clear instructions for
reproducing the problem. isolating and reproducing the problem.
See [How to Report Bugs Effectively](http://www.chiark.greenend.org.uk/~sgtatham/bugs.html). See [How to Report Bugs Effectively](http://www.chiark.greenend.org.uk/~sgtatham/bugs.html).
#### New features #### New features
Before submitting a pull request, please open an issue first to discuss it and Before submitting a pull request, please open an issue first to discuss it and
claim it. This prevents overlapping efforts and keeps the project in-line with claim it. This prevents overlapping efforts and keeps the project in-line with
its goals. If you prefer to discuss the feature privately, you can reach other its goals. If you prefer to discuss the feature privately, you can reach other
developers on Slack or you may email me directly. (My email address is below.) developers on Gitter or you may email me directly. (My email address is below.)
And don't forget to write tests for new features! And don't forget to write tests for new features!
......
[![Caddy](https://caddyserver.com/resources/images/caddy-boxed.png)](https://caddyserver.com) [![Caddy](https://caddyserver.com/resources/images/caddy-boxed.png)](https://caddyserver.com)
[![Dev Chat](https://img.shields.io/badge/dev%20chat-gitter-ff69b4.svg?style=flat-square&label=dev+chat&color=ff69b4)](https://gitter.im/mholt/caddy)
[![Documentation](https://img.shields.io/badge/godoc-reference-blue.svg?style=flat-square)](https://godoc.org/github.com/mholt/caddy) [![Documentation](https://img.shields.io/badge/godoc-reference-blue.svg?style=flat-square)](https://godoc.org/github.com/mholt/caddy)
[![Linux Build Status](https://img.shields.io/travis/mholt/caddy.svg?style=flat-square&label=linux+build)](https://travis-ci.org/mholt/caddy) [![Linux Build Status](https://img.shields.io/travis/mholt/caddy.svg?style=flat-square&label=linux+build)](https://travis-ci.org/mholt/caddy)
[![Windows Build Status](https://img.shields.io/appveyor/ci/mholt/caddy.svg?style=flat-square&label=windows+build)](https://ci.appveyor.com/project/mholt/caddy) [![Windows Build Status](https://img.shields.io/appveyor/ci/mholt/caddy.svg?style=flat-square&label=windows+build)](https://ci.appveyor.com/project/mholt/caddy)
...@@ -95,7 +96,7 @@ You may also be interested in the [developer guide] ...@@ -95,7 +96,7 @@ You may also be interested in the [developer guide]
## Running from Source ## Running from Source
Note: You will need **[Go 1.4](https://golang.org/dl)** or a later version. Note: You will need **[Go 1.4](https://golang.org/dl/)** or a later version.
1. `$ go get github.com/mholt/caddy` 1. `$ go get github.com/mholt/caddy`
2. `cd` into your website's directory 2. `cd` into your website's directory
...@@ -120,11 +121,11 @@ ports < 1024 like 80 and 443. ...@@ -120,11 +121,11 @@ ports < 1024 like 80 and 443.
Caddy is available as a Docker container from any of these sources: Caddy is available as a Docker container from any of these sources:
- [abiosoft/caddy](https://registry.hub.docker.com/u/abiosoft/caddy/) - [abiosoft/caddy](https://hub.docker.com/r/abiosoft/caddy/)
- [darron/caddy](https://registry.hub.docker.com/u/darron/caddy/) - [darron/caddy](https://hub.docker.com/r/darron/caddy/)
- [joshix/caddy](https://registry.hub.docker.com/u/joshix/caddy/) - [joshix/caddy](https://hub.docker.com/r/joshix/caddy/)
- [jumanjiman/caddy](https://registry.hub.docker.com/u/jumanjiman/caddy/) - [jumanjiman/caddy](https://hub.docker.com/r/jumanjiman/caddy/)
- [zenithar/nano-caddy](https://registry.hub.docker.com/u/zenithar/nano-caddy/) - [zenithar/nano-caddy](https://hub.docker.com/r/zenithar/nano-caddy/)
...@@ -139,9 +140,10 @@ packages that each Caddy package imports. ...@@ -139,9 +140,10 @@ packages that each Caddy package imports.
## Contributing ## Contributing
**[Join us on Slack](https://gophers.slack.com/messages/caddy/)** to chat with **[Join our dev chat on Gitter](https://gitter.im/mholt/caddy)** to chat with
other Caddy developers! ([Request an invite](http://bit.ly/go-slack-signup), other Caddy developers! (Dev chat only; try our
then join the #caddy channel.) [support room](https://gitter.im/caddyserver/support) for help or
[general](https://gitter.im/caddyserver/general) for anything else.)
This project would not be what it is without your help. Please see the This project would not be what it is without your help. Please see the
[contributing guidelines](https://github.com/mholt/caddy/blob/master/CONTRIBUTING.md) [contributing guidelines](https://github.com/mholt/caddy/blob/master/CONTRIBUTING.md)
......
...@@ -49,6 +49,7 @@ func ObtainCertsAndConfigure(configs []server.Config, optPort string) ([]server. ...@@ -49,6 +49,7 @@ func ObtainCertsAndConfigure(configs []server.Config, optPort string) ([]server.
// obtain certificates for configs that need one, and reconfigure each // obtain certificates for configs that need one, and reconfigure each
// config to use the certificates // config to use the certificates
finishedHosts := make(map[string]struct{})
for leEmail, cfgIndexes := range groupedConfigs { for leEmail, cfgIndexes := range groupedConfigs {
// make client to service this email address with CA server // make client to service this email address with CA server
client, err := newClientPort(leEmail, optPort) client, err := newClientPort(leEmail, optPort)
...@@ -60,6 +61,12 @@ func ObtainCertsAndConfigure(configs []server.Config, optPort string) ([]server. ...@@ -60,6 +61,12 @@ func ObtainCertsAndConfigure(configs []server.Config, optPort string) ([]server.
for _, idx := range cfgIndexes { for _, idx := range cfgIndexes {
hostname := configs[idx].Host hostname := configs[idx].Host
// prevent duplicate efforts, for example, when host is served on multiple ports
if _, ok := finishedHosts[hostname]; ok {
continue
}
finishedHosts[hostname] = struct{}{}
Obtain: Obtain:
certificate, failures := client.ObtainCertificate([]string{hostname}, true) certificate, failures := client.ObtainCertificate([]string{hostname}, true)
if len(failures) == 0 { if len(failures) == 0 {
......
...@@ -144,7 +144,7 @@ func getEmail(cfg server.Config) string { ...@@ -144,7 +144,7 @@ func getEmail(cfg server.Config) string {
// Alas, we must bother the user and ask for an email address; // Alas, we must bother the user and ask for an email address;
// if they proceed they also agree to the SA. // if they proceed they also agree to the SA.
reader := bufio.NewReader(stdin) reader := bufio.NewReader(stdin)
fmt.Println("Your sites will be served over HTTPS automatically using Let's Encrypt.") fmt.Println("\nYour sites will be served over HTTPS automatically using Let's Encrypt.")
fmt.Println("By continuing, you agree to the Let's Encrypt Subscriber Agreement at:") fmt.Println("By continuing, you agree to the Let's Encrypt Subscriber Agreement at:")
fmt.Println(" " + saURL) // TODO: Show current SA link fmt.Println(" " + saURL) // TODO: Show current SA link
fmt.Println("Please enter your email address so you can recover your account if needed.") fmt.Println("Please enter your email address so you can recover your account if needed.")
......
glob0.host0 {
dir2 arg1
}
glob0.host1 {
}
glob1.host0 {
dir1
dir2 arg1
}
glob2.host0 {
dir2 arg1
}
...@@ -177,19 +177,52 @@ func (p *parser) directives() error { ...@@ -177,19 +177,52 @@ func (p *parser) directives() error {
} }
// doImport swaps out the import directive and its argument // doImport swaps out the import directive and its argument
// (a total of 2 tokens) with the tokens in the file specified. // (a total of 2 tokens) with the tokens in the specified file
// When the function returns, the cursor is on the token before // or globbing pattern. When the function returns, the cursor
// where the import directive was. In other words, call Next() // is on the token before where the import directive was. In
// to access the first token that was imported. // other words, call Next() to access the first token that was
// imported.
func (p *parser) doImport() error { func (p *parser) doImport() error {
if !p.NextArg() { if !p.NextArg() {
return p.ArgErr() return p.ArgErr()
} }
importFile := p.Val() importPattern := p.Val()
if p.NextArg() { if p.NextArg() {
return p.Err("Import allows only one file to import") return p.Err("Import allows only one expression, either file or glob pattern")
} }
matches, err := filepath.Glob(importPattern)
if err != nil {
return p.Errf("Failed to use import pattern %s - %s", importPattern, err.Error())
}
if len(matches) == 0 {
return p.Errf("No files matching the import pattern %s", importPattern)
}
// Splice out the import directive and its argument (2 tokens total)
// and insert the imported tokens in their place.
tokensBefore := p.tokens[:p.cursor-1]
tokensAfter := p.tokens[p.cursor+1:]
// cursor was advanced one position to read filename; rewind it
p.cursor--
p.tokens = tokensBefore
for _, importFile := range matches {
if err := p.doSingleImport(importFile); err != nil {
return err
}
}
p.tokens = append(p.tokens, append(tokensAfter)...)
return nil
}
// doSingleImport lexes the individual files matching the
// globbing pattern from of the import directive.
func (p *parser) doSingleImport(importFile string) error {
file, err := os.Open(importFile) file, err := os.Open(importFile)
if err != nil { if err != nil {
return p.Errf("Could not import %s - %v", importFile, err) return p.Errf("Could not import %s - %v", importFile, err)
...@@ -204,10 +237,7 @@ func (p *parser) doImport() error { ...@@ -204,10 +237,7 @@ func (p *parser) doImport() error {
// Splice out the import directive and its argument (2 tokens total) // Splice out the import directive and its argument (2 tokens total)
// and insert the imported tokens in their place. // and insert the imported tokens in their place.
tokensBefore := p.tokens[:p.cursor-1] p.tokens = append(p.tokens, append(importedTokens)...)
tokensAfter := p.tokens[p.cursor+1:]
p.tokens = append(tokensBefore, append(importedTokens, tokensAfter...)...)
p.cursor-- // cursor was advanced one position to read the filename; rewind it
return nil return nil
} }
......
...@@ -333,6 +333,13 @@ func TestParseAll(t *testing.T) { ...@@ -333,6 +333,13 @@ func TestParseAll(t *testing.T) {
[]address{{"http://host1.com", "http", "host1.com", "80"}, {"http://host2.com", "http", "host2.com", "80"}}, []address{{"http://host1.com", "http", "host1.com", "80"}, {"http://host2.com", "http", "host2.com", "80"}},
[]address{{"https://host3.com", "https", "host3.com", "443"}, {"https://host4.com", "https", "host4.com", "443"}}, []address{{"https://host3.com", "https", "host3.com", "443"}, {"https://host4.com", "https", "host4.com", "443"}},
}}, }},
{`import import_glob*.txt`, false, [][]address{
[]address{{"glob0.host0", ""}},
[]address{{"glob0.host1", ""}},
[]address{{"glob1.host0", ""}},
[]address{{"glob2.host0", ""}},
}},
} { } {
p := testParser(test.input) p := testParser(test.input)
blocks, err := p.parseAll() blocks, err := p.parseAll()
......
...@@ -27,9 +27,13 @@ func gzipParse(c *Controller) ([]gzip.Config, error) { ...@@ -27,9 +27,13 @@ func gzipParse(c *Controller) ([]gzip.Config, error) {
for c.Next() { for c.Next() {
config := gzip.Config{} config := gzip.Config{}
// Request Filters
pathFilter := gzip.PathFilter{IgnoredPaths: make(gzip.Set)} pathFilter := gzip.PathFilter{IgnoredPaths: make(gzip.Set)}
extFilter := gzip.ExtFilter{Exts: make(gzip.Set)} extFilter := gzip.ExtFilter{Exts: make(gzip.Set)}
// Response Filters
lengthFilter := gzip.LengthFilter(0)
// No extra args expected // No extra args expected
if len(c.RemainingArgs()) > 0 { if len(c.RemainingArgs()) > 0 {
return configs, c.ArgErr() return configs, c.ArgErr()
...@@ -68,24 +72,42 @@ func gzipParse(c *Controller) ([]gzip.Config, error) { ...@@ -68,24 +72,42 @@ func gzipParse(c *Controller) ([]gzip.Config, error) {
} }
level, _ := strconv.Atoi(c.Val()) level, _ := strconv.Atoi(c.Val())
config.Level = level config.Level = level
case "min_length":
if !c.NextArg() {
return configs, c.ArgErr()
}
length, err := strconv.ParseInt(c.Val(), 10, 64)
if err != nil {
return configs, err
} else if length == 0 {
return configs, fmt.Errorf(`gzip: min_length must be greater than 0`)
}
lengthFilter = gzip.LengthFilter(length)
default: default:
return configs, c.ArgErr() return configs, c.ArgErr()
} }
} }
config.Filters = []gzip.Filter{} // Request Filters
config.RequestFilters = []gzip.RequestFilter{}
// If ignored paths are specified, put in front to filter with path first // If ignored paths are specified, put in front to filter with path first
if len(pathFilter.IgnoredPaths) > 0 { if len(pathFilter.IgnoredPaths) > 0 {
config.Filters = []gzip.Filter{pathFilter} config.RequestFilters = []gzip.RequestFilter{pathFilter}
} }
// Then, if extensions are specified, use those to filter. // Then, if extensions are specified, use those to filter.
// Otherwise, use default extensions filter. // Otherwise, use default extensions filter.
if len(extFilter.Exts) > 0 { if len(extFilter.Exts) > 0 {
config.Filters = append(config.Filters, extFilter) config.RequestFilters = append(config.RequestFilters, extFilter)
} else { } else {
config.Filters = append(config.Filters, gzip.DefaultExtFilter()) config.RequestFilters = append(config.RequestFilters, gzip.DefaultExtFilter())
}
// Response Filters
// If min_length is specified, use it.
if int64(lengthFilter) != 0 {
config.ResponseFilters = append(config.ResponseFilters, lengthFilter)
} }
configs = append(configs, config) configs = append(configs, config)
......
...@@ -73,6 +73,18 @@ func TestGzip(t *testing.T) { ...@@ -73,6 +73,18 @@ func TestGzip(t *testing.T) {
level 1 level 1
} }
`, false}, `, false},
{`gzip { not /file
ext *
level 1
min_length ab
}
`, true},
{`gzip { not /file
ext *
level 1
min_length 1000
}
`, false},
} }
for i, test := range tests { for i, test := range tests {
c := NewTestController(test.input) c := NewTestController(test.input)
......
package setup package setup
import ( import (
"net/http"
"strings"
"github.com/mholt/caddy/middleware" "github.com/mholt/caddy/middleware"
"github.com/mholt/caddy/middleware/rewrite" "github.com/mholt/caddy/middleware/rewrite"
) )
...@@ -13,7 +16,11 @@ func Rewrite(c *Controller) (middleware.Middleware, error) { ...@@ -13,7 +16,11 @@ func Rewrite(c *Controller) (middleware.Middleware, error) {
} }
return func(next middleware.Handler) middleware.Handler { return func(next middleware.Handler) middleware.Handler {
return rewrite.Rewrite{Next: next, Rules: rewrites} return rewrite.Rewrite{
Next: next,
FileSys: http.Dir(c.Root),
Rules: rewrites,
}
}, nil }, nil
} }
...@@ -30,6 +37,8 @@ func rewriteParse(c *Controller) ([]rewrite.Rule, error) { ...@@ -30,6 +37,8 @@ func rewriteParse(c *Controller) ([]rewrite.Rule, error) {
args := c.RemainingArgs() args := c.RemainingArgs()
var ifs []rewrite.If
switch len(args) { switch len(args) {
case 2: case 2:
rule = rewrite.NewSimpleRule(args[0], args[1]) rule = rewrite.NewSimpleRule(args[0], args[1])
...@@ -46,25 +55,36 @@ func rewriteParse(c *Controller) ([]rewrite.Rule, error) { ...@@ -46,25 +55,36 @@ func rewriteParse(c *Controller) ([]rewrite.Rule, error) {
} }
pattern = c.Val() pattern = c.Val()
case "to": case "to":
if !c.NextArg() { args1 := c.RemainingArgs()
if len(args1) == 0 {
return nil, c.ArgErr() return nil, c.ArgErr()
} }
to = c.Val() to = strings.Join(args1, " ")
case "ext": case "ext":
args1 := c.RemainingArgs() args1 := c.RemainingArgs()
if len(args1) == 0 { if len(args1) == 0 {
return nil, c.ArgErr() return nil, c.ArgErr()
} }
ext = args1 ext = args1
case "if":
args1 := c.RemainingArgs()
if len(args1) != 3 {
return nil, c.ArgErr()
}
ifCond, err := rewrite.NewIf(args1[0], args1[1], args1[2])
if err != nil {
return nil, err
}
ifs = append(ifs, ifCond)
default: default:
return nil, c.ArgErr() return nil, c.ArgErr()
} }
} }
// ensure pattern and to are specified // ensure to is specified
if pattern == "" || to == "" { if to == "" {
return nil, c.ArgErr() return nil, c.ArgErr()
} }
if rule, err = rewrite.NewRegexpRule(base, pattern, to, ext); err != nil { if rule, err = rewrite.NewComplexRule(base, pattern, to, ext, ifs); err != nil {
return nil, err return nil, err
} }
regexpRules = append(regexpRules, rule) regexpRules = append(regexpRules, rule)
......
package setup package setup
import ( import (
"testing"
"fmt" "fmt"
"regexp" "regexp"
"testing"
"github.com/mholt/caddy/middleware/rewrite" "github.com/mholt/caddy/middleware/rewrite"
) )
...@@ -96,16 +95,16 @@ func TestRewriteParse(t *testing.T) { ...@@ -96,16 +95,16 @@ func TestRewriteParse(t *testing.T) {
}{ }{
{`rewrite { {`rewrite {
r .* r .*
to /to to /to /index.php?
}`, false, []rewrite.Rule{ }`, false, []rewrite.Rule{
&rewrite.RegexpRule{Base: "/", To: "/to", Regexp: regexp.MustCompile(".*")}, &rewrite.ComplexRule{Base: "/", To: "/to /index.php?", Regexp: regexp.MustCompile(".*")},
}}, }},
{`rewrite { {`rewrite {
regexp .* regexp .*
to /to to /to
ext / html txt ext / html txt
}`, false, []rewrite.Rule{ }`, false, []rewrite.Rule{
&rewrite.RegexpRule{Base: "/", To: "/to", Exts: []string{"/", "html", "txt"}, Regexp: regexp.MustCompile(".*")}, &rewrite.ComplexRule{Base: "/", To: "/to", Exts: []string{"/", "html", "txt"}, Regexp: regexp.MustCompile(".*")},
}}, }},
{`rewrite /path { {`rewrite /path {
r rr r rr
...@@ -113,29 +112,30 @@ func TestRewriteParse(t *testing.T) { ...@@ -113,29 +112,30 @@ func TestRewriteParse(t *testing.T) {
} }
rewrite / { rewrite / {
regexp [a-z]+ regexp [a-z]+
to /to to /to /to2
} }
`, false, []rewrite.Rule{ `, false, []rewrite.Rule{
&rewrite.RegexpRule{Base: "/path", To: "/dest", Regexp: regexp.MustCompile("rr")}, &rewrite.ComplexRule{Base: "/path", To: "/dest", Regexp: regexp.MustCompile("rr")},
&rewrite.RegexpRule{Base: "/", To: "/to", Regexp: regexp.MustCompile("[a-z]+")}, &rewrite.ComplexRule{Base: "/", To: "/to /to2", Regexp: regexp.MustCompile("[a-z]+")},
}},
{`rewrite {
to /to
}`, true, []rewrite.Rule{
&rewrite.RegexpRule{},
}}, }},
{`rewrite { {`rewrite {
r .* r .*
}`, true, []rewrite.Rule{ }`, true, []rewrite.Rule{
&rewrite.RegexpRule{}, &rewrite.ComplexRule{},
}}, }},
{`rewrite { {`rewrite {
}`, true, []rewrite.Rule{ }`, true, []rewrite.Rule{
&rewrite.RegexpRule{}, &rewrite.ComplexRule{},
}}, }},
{`rewrite /`, true, []rewrite.Rule{ {`rewrite /`, true, []rewrite.Rule{
&rewrite.RegexpRule{}, &rewrite.ComplexRule{},
}},
{`rewrite {
to /to
if {path} is a
}`, false, []rewrite.Rule{
&rewrite.ComplexRule{Base: "/", To: "/to", Ifs: []rewrite.If{rewrite.If{A: "{path}", Operator: "is", B: "a"}}},
}}, }},
} }
...@@ -157,8 +157,8 @@ func TestRewriteParse(t *testing.T) { ...@@ -157,8 +157,8 @@ func TestRewriteParse(t *testing.T) {
} }
for j, e := range test.expected { for j, e := range test.expected {
actualRule := actual[j].(*rewrite.RegexpRule) actualRule := actual[j].(*rewrite.ComplexRule)
expectedRule := e.(*rewrite.RegexpRule) expectedRule := e.(*rewrite.ComplexRule)
if actualRule.Base != expectedRule.Base { if actualRule.Base != expectedRule.Base {
t.Errorf("Test %d, rule %d: Expected Base=%s, got %s", t.Errorf("Test %d, rule %d: Expected Base=%s, got %s",
...@@ -175,11 +175,19 @@ func TestRewriteParse(t *testing.T) { ...@@ -175,11 +175,19 @@ func TestRewriteParse(t *testing.T) {
i, j, expectedRule.To, actualRule.To) i, j, expectedRule.To, actualRule.To)
} }
if actualRule.Regexp != nil {
if actualRule.String() != expectedRule.String() { if actualRule.String() != expectedRule.String() {
t.Errorf("Test %d, rule %d: Expected Pattern=%s, got %s", t.Errorf("Test %d, rule %d: Expected Pattern=%s, got %s",
i, j, expectedRule.String(), actualRule.String()) i, j, expectedRule.String(), actualRule.String())
} }
} }
if fmt.Sprint(actualRule.Ifs) != fmt.Sprint(expectedRule.Ifs) {
t.Errorf("Test %d, rule %d: Expected Pattern=%s, got %s",
i, j, fmt.Sprint(expectedRule.Ifs), fmt.Sprint(actualRule.Ifs))
}
}
} }
} }
...@@ -91,7 +91,7 @@ func TLS(c *Controller) (middleware.Middleware, error) { ...@@ -91,7 +91,7 @@ func TLS(c *Controller) (middleware.Middleware, error) {
func SetDefaultTLSParams(c *server.Config) { func SetDefaultTLSParams(c *server.Config) {
// If no ciphers provided, use all that Caddy supports for the protocol // If no ciphers provided, use all that Caddy supports for the protocol
if len(c.TLS.Ciphers) == 0 { if len(c.TLS.Ciphers) == 0 {
c.TLS.Ciphers = supportedCiphers c.TLS.Ciphers = defaultCiphers
} }
// Not a cipher suite, but still important for mitigating protocol downgrade attacks // Not a cipher suite, but still important for mitigating protocol downgrade attacks
...@@ -160,3 +160,15 @@ var supportedCiphers = []uint16{ ...@@ -160,3 +160,15 @@ var supportedCiphers = []uint16{
tls.TLS_ECDHE_RSA_WITH_3DES_EDE_CBC_SHA, tls.TLS_ECDHE_RSA_WITH_3DES_EDE_CBC_SHA,
tls.TLS_RSA_WITH_3DES_EDE_CBC_SHA, tls.TLS_RSA_WITH_3DES_EDE_CBC_SHA,
} }
// List of all the ciphers we want to use by default
var defaultCiphers = []uint16{
tls.TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,
tls.TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,
tls.TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA,
tls.TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA,
tls.TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA,
tls.TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA,
tls.TLS_RSA_WITH_AES_256_CBC_SHA,
tls.TLS_RSA_WITH_AES_128_CBC_SHA,
}
...@@ -42,15 +42,13 @@ func TestTLSParseBasic(t *testing.T) { ...@@ -42,15 +42,13 @@ func TestTLSParseBasic(t *testing.T) {
tls.TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA, tls.TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA,
tls.TLS_RSA_WITH_AES_256_CBC_SHA, tls.TLS_RSA_WITH_AES_256_CBC_SHA,
tls.TLS_RSA_WITH_AES_128_CBC_SHA, tls.TLS_RSA_WITH_AES_128_CBC_SHA,
tls.TLS_ECDHE_RSA_WITH_3DES_EDE_CBC_SHA,
tls.TLS_RSA_WITH_3DES_EDE_CBC_SHA,
tls.TLS_FALLBACK_SCSV, tls.TLS_FALLBACK_SCSV,
} }
// Ensure count is correct (plus one for TLS_FALLBACK_SCSV) // Ensure count is correct (plus one for TLS_FALLBACK_SCSV)
if len(c.TLS.Ciphers) != len(supportedCiphers)+1 { if len(c.TLS.Ciphers) != len(expectedCiphers) {
t.Errorf("Expected %v Ciphers (including TLS_FALLBACK_SCSV), got %v", t.Errorf("Expected %v Ciphers (including TLS_FALLBACK_SCSV), got %v",
len(supportedCiphers)+1, len(c.TLS.Ciphers)) len(expectedCiphers), len(c.TLS.Ciphers))
} }
// Ensure ordering is correct // Ensure ordering is correct
......
CHANGES CHANGES
0.8 beta 0.8.0 (December 4, 2015)
- Let's Encrypt (free, automatic, fully-managed HTTPS for your sites) - HTTPS by default via Let's Encrypt (certs & keys are fully managed)
- Graceful restarts (for POSIX-compatible systems) - Graceful restarts (on POSIX-compliant systems)
- Major internal refactoring to allow use of Caddy as library - Major internal refactoring to allow use of Caddy as library
- New directive 'mime' to customize Content-Type based on file extension - New directive 'mime' to customize Content-Type based on file extension
- New -accept flag to accept Let's Encrypt SA without prompt - New -accept flag to accept Let's Encrypt SA without prompt
...@@ -14,6 +14,7 @@ CHANGES ...@@ -14,6 +14,7 @@ CHANGES
- New -grace flag to customize the graceful shutdown timeout - New -grace flag to customize the graceful shutdown timeout
- New support for SIGHUP, SIGTERM, and SIGQUIT signals - New support for SIGHUP, SIGTERM, and SIGQUIT signals
- browse: Render filenames with multiple whitespace properly - browse: Render filenames with multiple whitespace properly
- core: Use environment variables in Caddyfile
- markdown: Include Last-Modified header in response - markdown: Include Last-Modified header in response
- markdown: Render tables, strikethrough, and fenced code blocks - markdown: Render tables, strikethrough, and fenced code blocks
- proxy: Ability to exclude/ignore paths from proxying - proxy: Ability to exclude/ignore paths from proxying
......
CADDY 0.8 beta 4 CADDY 0.8
Website Website
https://caddyserver.com https://caddyserver.com
......
...@@ -27,7 +27,7 @@ var ( ...@@ -27,7 +27,7 @@ var (
const ( const (
appName = "Caddy" appName = "Caddy"
appVersion = "0.8 beta 4" appVersion = "0.8"
) )
func init() { func init() {
......
...@@ -115,7 +115,13 @@ func (h Handler) ServeHTTP(w http.ResponseWriter, r *http.Request) (int, error) ...@@ -115,7 +115,13 @@ func (h Handler) ServeHTTP(w http.ResponseWriter, r *http.Request) (int, error)
return http.StatusBadGateway, err return http.StatusBadGateway, err
} }
return 0, nil // FastCGI stderr outputs
if fcgi.stderr.Len() != 0 {
// Remove trailing newline, error logger already does this.
err = LogError(strings.TrimSuffix(fcgi.stderr.String(), "\n"))
}
return resp.StatusCode, err
} }
} }
...@@ -281,3 +287,11 @@ var ( ...@@ -281,3 +287,11 @@ var (
// ErrIndexMissingSplit describes an index configuration error. // ErrIndexMissingSplit describes an index configuration error.
ErrIndexMissingSplit = errors.New("configured index file(s) must include split value") ErrIndexMissingSplit = errors.New("configured index file(s) must include split value")
) )
// LogError is a non fatal error that allows requests to go through.
type LogError string
// Error satisfies error interface.
func (l LogError) Error() string {
return string(l)
}
...@@ -164,6 +164,7 @@ type FCGIClient struct { ...@@ -164,6 +164,7 @@ type FCGIClient struct {
rwc io.ReadWriteCloser rwc io.ReadWriteCloser
h header h header
buf bytes.Buffer buf bytes.Buffer
stderr bytes.Buffer
keepAlive bool keepAlive bool
reqID uint16 reqID uint16
} }
...@@ -346,11 +347,23 @@ func (w *streamReader) Read(p []byte) (n int, err error) { ...@@ -346,11 +347,23 @@ func (w *streamReader) Read(p []byte) (n int, err error) {
if len(p) > 0 { if len(p) > 0 {
if len(w.buf) == 0 { if len(w.buf) == 0 {
// filter outputs for error log
for {
rec := &record{} rec := &record{}
w.buf, err = rec.read(w.c.rwc) var buf []byte
buf, err = rec.read(w.c.rwc)
if err != nil { if err != nil {
return return
} }
// standard error output
if rec.h.Type == Stderr {
w.c.stderr.Write(buf)
continue
}
w.buf = buf
break
}
} }
n = len(p) n = len(p)
...@@ -387,6 +400,15 @@ func (c *FCGIClient) Do(p map[string]string, req io.Reader) (r io.Reader, err er ...@@ -387,6 +400,15 @@ func (c *FCGIClient) Do(p map[string]string, req io.Reader) (r io.Reader, err er
return return
} }
// clientCloser is a io.ReadCloser. It wraps a io.Reader with a Closer
// that closes FCGIClient connection.
type clientCloser struct {
*FCGIClient
io.Reader
}
func (f clientCloser) Close() error { return f.rwc.Close() }
// Request returns a HTTP Response with Header and Body // Request returns a HTTP Response with Header and Body
// from fcgi responder // from fcgi responder
func (c *FCGIClient) Request(p map[string]string, req io.Reader) (resp *http.Response, err error) { func (c *FCGIClient) Request(p map[string]string, req io.Reader) (resp *http.Response, err error) {
...@@ -426,9 +448,9 @@ func (c *FCGIClient) Request(p map[string]string, req io.Reader) (resp *http.Res ...@@ -426,9 +448,9 @@ func (c *FCGIClient) Request(p map[string]string, req io.Reader) (resp *http.Res
resp.ContentLength, _ = strconv.ParseInt(resp.Header.Get("Content-Length"), 10, 64) resp.ContentLength, _ = strconv.ParseInt(resp.Header.Get("Content-Length"), 10, 64)
if chunked(resp.TransferEncoding) { if chunked(resp.TransferEncoding) {
resp.Body = ioutil.NopCloser(httputil.NewChunkedReader(rb)) resp.Body = clientCloser{c, httputil.NewChunkedReader(rb)}
} else { } else {
resp.Body = ioutil.NopCloser(rb) resp.Body = clientCloser{c, ioutil.NopCloser(rb)}
} }
return return
} }
......
...@@ -6,6 +6,7 @@ import ( ...@@ -6,6 +6,7 @@ import (
"compress/gzip" "compress/gzip"
"fmt" "fmt"
"io" "io"
"io/ioutil"
"net/http" "net/http"
"strings" "strings"
...@@ -23,7 +24,8 @@ type Gzip struct { ...@@ -23,7 +24,8 @@ type Gzip struct {
// Config holds the configuration for Gzip middleware // Config holds the configuration for Gzip middleware
type Config struct { type Config struct {
Filters []Filter // Filters to use RequestFilters []RequestFilter
ResponseFilters []ResponseFilter
Level int // Compression level Level int // Compression level
} }
...@@ -36,8 +38,8 @@ func (g Gzip) ServeHTTP(w http.ResponseWriter, r *http.Request) (int, error) { ...@@ -36,8 +38,8 @@ func (g Gzip) ServeHTTP(w http.ResponseWriter, r *http.Request) (int, error) {
outer: outer:
for _, c := range g.Configs { for _, c := range g.Configs {
// Check filters to determine if gzipping is permitted for this request // Check request filters to determine if gzipping is permitted for this request
for _, filter := range c.Filters { for _, filter := range c.RequestFilters {
if !filter.ShouldCompress(r) { if !filter.ShouldCompress(r) {
continue outer continue outer
} }
...@@ -46,18 +48,30 @@ outer: ...@@ -46,18 +48,30 @@ outer:
// Delete this header so gzipping is not repeated later in the chain // Delete this header so gzipping is not repeated later in the chain
r.Header.Del("Accept-Encoding") r.Header.Del("Accept-Encoding")
w.Header().Set("Content-Encoding", "gzip") // gzipWriter modifies underlying writer at init,
w.Header().Set("Vary", "Accept-Encoding") // use a discard writer instead to leave ResponseWriter in
gzipWriter, err := newWriter(c, w) // original form.
gzipWriter, err := newWriter(c, ioutil.Discard)
if err != nil { if err != nil {
// should not happen // should not happen
return http.StatusInternalServerError, err return http.StatusInternalServerError, err
} }
defer gzipWriter.Close() defer gzipWriter.Close()
gz := gzipResponseWriter{Writer: gzipWriter, ResponseWriter: w} gz := &gzipResponseWriter{Writer: gzipWriter, ResponseWriter: w}
var rw http.ResponseWriter
// if no response filter is used
if len(c.ResponseFilters) == 0 {
// replace discard writer with ResponseWriter
gzipWriter.Reset(w)
rw = gz
} else {
// wrap gzip writer with ResponseFilterWriter
rw = NewResponseFilterWriter(c.ResponseFilters, gz)
}
// Any response in forward middleware will now be compressed // Any response in forward middleware will now be compressed
status, err := g.Next.ServeHTTP(gz, r) status, err := g.Next.ServeHTTP(rw, r)
// If there was an error that remained unhandled, we need // If there was an error that remained unhandled, we need
// to send something back before gzipWriter gets closed at // to send something back before gzipWriter gets closed at
...@@ -78,7 +92,7 @@ outer: ...@@ -78,7 +92,7 @@ outer:
// newWriter create a new Gzip Writer based on the compression level. // newWriter create a new Gzip Writer based on the compression level.
// If the level is valid (i.e. between 1 and 9), it uses the level. // If the level is valid (i.e. between 1 and 9), it uses the level.
// Otherwise, it uses default compression level. // Otherwise, it uses default compression level.
func newWriter(c Config, w http.ResponseWriter) (*gzip.Writer, error) { func newWriter(c Config, w io.Writer) (*gzip.Writer, error) {
if c.Level >= gzip.BestSpeed && c.Level <= gzip.BestCompression { if c.Level >= gzip.BestSpeed && c.Level <= gzip.BestCompression {
return gzip.NewWriterLevel(w, c.Level) return gzip.NewWriterLevel(w, c.Level)
} }
...@@ -90,19 +104,26 @@ func newWriter(c Config, w http.ResponseWriter) (*gzip.Writer, error) { ...@@ -90,19 +104,26 @@ func newWriter(c Config, w http.ResponseWriter) (*gzip.Writer, error) {
type gzipResponseWriter struct { type gzipResponseWriter struct {
io.Writer io.Writer
http.ResponseWriter http.ResponseWriter
statusCodeWritten bool
} }
// WriteHeader wraps the underlying WriteHeader method to prevent // WriteHeader wraps the underlying WriteHeader method to prevent
// problems with conflicting headers from proxied backends. For // problems with conflicting headers from proxied backends. For
// example, a backend system that calculates Content-Length would // example, a backend system that calculates Content-Length would
// be wrong because it doesn't know it's being gzipped. // be wrong because it doesn't know it's being gzipped.
func (w gzipResponseWriter) WriteHeader(code int) { func (w *gzipResponseWriter) WriteHeader(code int) {
w.Header().Del("Content-Length") w.Header().Del("Content-Length")
w.Header().Set("Content-Encoding", "gzip")
w.Header().Add("Vary", "Accept-Encoding")
w.ResponseWriter.WriteHeader(code) w.ResponseWriter.WriteHeader(code)
w.statusCodeWritten = true
} }
// Write wraps the underlying Write method to do compression. // Write wraps the underlying Write method to do compression.
func (w gzipResponseWriter) Write(b []byte) (int, error) { func (w *gzipResponseWriter) Write(b []byte) (int, error) {
if !w.statusCodeWritten {
w.WriteHeader(http.StatusOK)
}
if w.Header().Get("Content-Type") == "" { if w.Header().Get("Content-Type") == "" {
w.Header().Set("Content-Type", http.DetectContentType(b)) w.Header().Set("Content-Type", http.DetectContentType(b))
} }
......
...@@ -21,7 +21,7 @@ func TestGzipHandler(t *testing.T) { ...@@ -21,7 +21,7 @@ func TestGzipHandler(t *testing.T) {
extFilter.Exts.Add(e) extFilter.Exts.Add(e)
} }
gz := Gzip{Configs: []Config{ gz := Gzip{Configs: []Config{
{Filters: []Filter{pathFilter, extFilter}}, {RequestFilters: []RequestFilter{pathFilter, extFilter}},
}} }}
w := httptest.NewRecorder() w := httptest.NewRecorder()
...@@ -80,6 +80,7 @@ func TestGzipHandler(t *testing.T) { ...@@ -80,6 +80,7 @@ func TestGzipHandler(t *testing.T) {
func nextFunc(shouldGzip bool) middleware.Handler { func nextFunc(shouldGzip bool) middleware.Handler {
return middleware.HandlerFunc(func(w http.ResponseWriter, r *http.Request) (int, error) { return middleware.HandlerFunc(func(w http.ResponseWriter, r *http.Request) (int, error) {
w.Write([]byte("test"))
if shouldGzip { if shouldGzip {
if r.Header.Get("Accept-Encoding") != "" { if r.Header.Get("Accept-Encoding") != "" {
return 0, fmt.Errorf("Accept-Encoding header not expected") return 0, fmt.Errorf("Accept-Encoding header not expected")
...@@ -90,7 +91,7 @@ func nextFunc(shouldGzip bool) middleware.Handler { ...@@ -90,7 +91,7 @@ func nextFunc(shouldGzip bool) middleware.Handler {
if w.Header().Get("Vary") != "Accept-Encoding" { if w.Header().Get("Vary") != "Accept-Encoding" {
return 0, fmt.Errorf("Vary must be Accept-Encoding, found %v", r.Header.Get("Vary")) return 0, fmt.Errorf("Vary must be Accept-Encoding, found %v", r.Header.Get("Vary"))
} }
if _, ok := w.(gzipResponseWriter); !ok { if _, ok := w.(*gzipResponseWriter); !ok {
return 0, fmt.Errorf("ResponseWriter should be gzipResponseWriter, found %T", w) return 0, fmt.Errorf("ResponseWriter should be gzipResponseWriter, found %T", w)
} }
return 0, nil return 0, nil
...@@ -101,7 +102,7 @@ func nextFunc(shouldGzip bool) middleware.Handler { ...@@ -101,7 +102,7 @@ func nextFunc(shouldGzip bool) middleware.Handler {
if w.Header().Get("Content-Encoding") == "gzip" { if w.Header().Get("Content-Encoding") == "gzip" {
return 0, fmt.Errorf("Content-Encoding must not be gzip, found gzip") return 0, fmt.Errorf("Content-Encoding must not be gzip, found gzip")
} }
if _, ok := w.(gzipResponseWriter); ok { if _, ok := w.(*gzipResponseWriter); ok {
return 0, fmt.Errorf("ResponseWriter should not be gzipResponseWriter") return 0, fmt.Errorf("ResponseWriter should not be gzipResponseWriter")
} }
return 0, nil return 0, nil
......
...@@ -7,8 +7,8 @@ import ( ...@@ -7,8 +7,8 @@ import (
"github.com/mholt/caddy/middleware" "github.com/mholt/caddy/middleware"
) )
// Filter determines if a request should be gzipped. // RequestFilter determines if a request should be gzipped.
type Filter interface { type RequestFilter interface {
// ShouldCompress tells if gzip compression // ShouldCompress tells if gzip compression
// should be done on the request. // should be done on the request.
ShouldCompress(*http.Request) bool ShouldCompress(*http.Request) bool
...@@ -26,7 +26,7 @@ func DefaultExtFilter() ExtFilter { ...@@ -26,7 +26,7 @@ func DefaultExtFilter() ExtFilter {
return m return m
} }
// ExtFilter is Filter for file name extensions. // ExtFilter is RequestFilter for file name extensions.
type ExtFilter struct { type ExtFilter struct {
// Exts is the file name extensions to accept // Exts is the file name extensions to accept
Exts Set Exts Set
...@@ -43,7 +43,7 @@ func (e ExtFilter) ShouldCompress(r *http.Request) bool { ...@@ -43,7 +43,7 @@ func (e ExtFilter) ShouldCompress(r *http.Request) bool {
return e.Exts.Contains(ExtWildCard) || e.Exts.Contains(ext) return e.Exts.Contains(ExtWildCard) || e.Exts.Contains(ext)
} }
// PathFilter is Filter for request path. // PathFilter is RequestFilter for request path.
type PathFilter struct { type PathFilter struct {
// IgnoredPaths is the paths to ignore // IgnoredPaths is the paths to ignore
IgnoredPaths Set IgnoredPaths Set
......
...@@ -47,7 +47,7 @@ func TestSet(t *testing.T) { ...@@ -47,7 +47,7 @@ func TestSet(t *testing.T) {
} }
func TestExtFilter(t *testing.T) { func TestExtFilter(t *testing.T) {
var filter Filter = ExtFilter{make(Set)} var filter RequestFilter = ExtFilter{make(Set)}
for _, e := range []string{".txt", ".html", ".css", ".md"} { for _, e := range []string{".txt", ".html", ".css", ".md"} {
filter.(ExtFilter).Exts.Add(e) filter.(ExtFilter).Exts.Add(e)
} }
...@@ -86,7 +86,7 @@ func TestPathFilter(t *testing.T) { ...@@ -86,7 +86,7 @@ func TestPathFilter(t *testing.T) {
paths := []string{ paths := []string{
"/a", "/b", "/c", "/de", "/a", "/b", "/c", "/de",
} }
var filter Filter = PathFilter{make(Set)} var filter RequestFilter = PathFilter{make(Set)}
for _, p := range paths { for _, p := range paths {
filter.(PathFilter).IgnoredPaths.Add(p) filter.(PathFilter).IgnoredPaths.Add(p)
} }
......
package gzip
import (
"compress/gzip"
"net/http"
"strconv"
)
// ResponseFilter determines if the response should be gzipped.
type ResponseFilter interface {
ShouldCompress(http.ResponseWriter) bool
}
// LengthFilter is ResponseFilter for minimum content length.
type LengthFilter int64
// ShouldCompress returns if content length is greater than or
// equals to minimum length.
func (l LengthFilter) ShouldCompress(w http.ResponseWriter) bool {
contentLength := w.Header().Get("Content-Length")
length, err := strconv.ParseInt(contentLength, 10, 64)
if err != nil || length == 0 {
return false
}
return l != 0 && int64(l) <= length
}
// ResponseFilterWriter validates ResponseFilters. It writes
// gzip compressed data if ResponseFilters are satisfied or
// uncompressed data otherwise.
type ResponseFilterWriter struct {
filters []ResponseFilter
shouldCompress bool
statusCodeWritten bool
*gzipResponseWriter
}
// NewResponseFilterWriter creates and initializes a new ResponseFilterWriter.
func NewResponseFilterWriter(filters []ResponseFilter, gz *gzipResponseWriter) *ResponseFilterWriter {
return &ResponseFilterWriter{filters: filters, gzipResponseWriter: gz}
}
// Write wraps underlying WriteHeader method and compresses if filters
// are satisfied.
func (r *ResponseFilterWriter) WriteHeader(code int) {
// Determine if compression should be used or not.
r.shouldCompress = true
for _, filter := range r.filters {
if !filter.ShouldCompress(r) {
r.shouldCompress = false
break
}
}
if r.shouldCompress {
// replace discard writer with ResponseWriter
if gzWriter, ok := r.gzipResponseWriter.Writer.(*gzip.Writer); ok {
gzWriter.Reset(r.ResponseWriter)
}
// use gzip WriteHeader to include and delete
// necessary headers
r.gzipResponseWriter.WriteHeader(code)
} else {
r.ResponseWriter.WriteHeader(code)
}
r.statusCodeWritten = true
}
// Write wraps underlying Write method and compresses if filters
// are satisfied
func (r *ResponseFilterWriter) Write(b []byte) (int, error) {
if !r.statusCodeWritten {
r.WriteHeader(http.StatusOK)
}
if r.shouldCompress {
return r.gzipResponseWriter.Write(b)
}
return r.ResponseWriter.Write(b)
}
package gzip
import (
"compress/gzip"
"fmt"
"net/http"
"net/http/httptest"
"testing"
"github.com/mholt/caddy/middleware"
)
func TestLengthFilter(t *testing.T) {
var filters []ResponseFilter = []ResponseFilter{
LengthFilter(100),
LengthFilter(1000),
LengthFilter(0),
}
var tests = []struct {
length int64
shouldCompress [3]bool
}{
{20, [3]bool{false, false, false}},
{50, [3]bool{false, false, false}},
{100, [3]bool{true, false, false}},
{500, [3]bool{true, false, false}},
{1000, [3]bool{true, true, false}},
{1500, [3]bool{true, true, false}},
}
for i, ts := range tests {
for j, filter := range filters {
r := httptest.NewRecorder()
r.Header().Set("Content-Length", fmt.Sprint(ts.length))
wWriter := NewResponseFilterWriter([]ResponseFilter{filter}, &gzipResponseWriter{gzip.NewWriter(r), r, false})
if filter.ShouldCompress(wWriter) != ts.shouldCompress[j] {
t.Errorf("Test %v: Expected %v found %v", i, ts.shouldCompress[j], filter.ShouldCompress(r))
}
}
}
}
func TestResponseFilterWriter(t *testing.T) {
tests := []struct {
body string
shouldCompress bool
}{
{"Hello\t\t\t\n", false},
{"Hello the \t\t\t world is\n\n\n great", true},
{"Hello \t\t\nfrom gzip", true},
{"Hello gzip\n", false},
}
filters := []ResponseFilter{
LengthFilter(15),
}
server := Gzip{Configs: []Config{
{ResponseFilters: filters},
}}
for i, ts := range tests {
server.Next = middleware.HandlerFunc(func(w http.ResponseWriter, r *http.Request) (int, error) {
w.Header().Set("Content-Length", fmt.Sprint(len(ts.body)))
w.Write([]byte(ts.body))
return 200, nil
})
r := urlRequest("/")
r.Header.Set("Accept-Encoding", "gzip")
w := httptest.NewRecorder()
server.ServeHTTP(w, r)
resp := w.Body.String()
if !ts.shouldCompress {
if resp != ts.body {
t.Errorf("Test %v: No compression expected, found %v", i, resp)
}
} else {
if resp == ts.body {
t.Errorf("Test %v: Compression expected, found %v", i, resp)
}
}
}
}
...@@ -5,6 +5,8 @@ import ( ...@@ -5,6 +5,8 @@ import (
"encoding/hex" "encoding/hex"
"fmt" "fmt"
"io/ioutil" "io/ioutil"
"net/http"
"net/url"
"os" "os"
"path/filepath" "path/filepath"
"strings" "strings"
...@@ -103,8 +105,12 @@ func generateStaticHTML(md Markdown, cfg *Config) error { ...@@ -103,8 +105,12 @@ func generateStaticHTML(md Markdown, cfg *Config) error {
reqPath = filepath.ToSlash(reqPath) reqPath = filepath.ToSlash(reqPath)
reqPath = "/" + reqPath reqPath = "/" + reqPath
// Create empty requests and url to cater for template values.
req, _ := http.NewRequest("", "/", nil)
urlVar, _ := url.Parse("/")
// Generate the static file // Generate the static file
ctx := middleware.Context{Root: md.FileSys} ctx := middleware.Context{Root: md.FileSys, Req: req, URL: urlVar}
_, err = md.Process(cfg, reqPath, body, ctx) _, err = md.Process(cfg, reqPath, body, ctx)
if err != nil { if err != nil {
return err return err
......
package markdown package markdown
import ( import (
"bufio"
"bytes" "bytes"
"encoding/json" "encoding/json"
"fmt" "fmt"
"io" "time"
"github.com/BurntSushi/toml" "github.com/BurntSushi/toml"
"gopkg.in/yaml.v2" "gopkg.in/yaml.v2"
"time"
) )
// Metadata stores a page's metadata // Metadata stores a page's metadata
...@@ -73,23 +71,20 @@ type JSONMetadataParser struct { ...@@ -73,23 +71,20 @@ type JSONMetadataParser struct {
// Parse the metadata // Parse the metadata
func (j *JSONMetadataParser) Parse(b []byte) ([]byte, error) { func (j *JSONMetadataParser) Parse(b []byte) ([]byte, error) {
b, markdown, err := extractMetadata(j, b)
if err != nil {
return markdown, err
}
m := make(map[string]interface{}) m := make(map[string]interface{})
// Read the preceding JSON object // Read the preceding JSON object
decoder := json.NewDecoder(bytes.NewReader(b)) decoder := json.NewDecoder(bytes.NewReader(b))
if err := decoder.Decode(&m); err != nil { if err := decoder.Decode(&m); err != nil {
return b, err return markdown, err
} }
j.metadata.load(m) j.metadata.load(m)
// Retrieve remaining bytes after decoding return markdown, nil
buf := make([]byte, len(b))
n, err := decoder.Buffered().Read(buf)
if err != nil {
return b, err
}
return buf[:n], nil
} }
// Metadata returns parsed metadata. It should be called // Metadata returns parsed metadata. It should be called
...@@ -183,43 +178,29 @@ func (y *YAMLMetadataParser) Closing() []byte { ...@@ -183,43 +178,29 @@ func (y *YAMLMetadataParser) Closing() []byte {
// It returns the metadata, the remaining bytes (markdown), and an error, if any. // It returns the metadata, the remaining bytes (markdown), and an error, if any.
func extractMetadata(parser MetadataParser, b []byte) (metadata []byte, markdown []byte, err error) { func extractMetadata(parser MetadataParser, b []byte) (metadata []byte, markdown []byte, err error) {
b = bytes.TrimSpace(b) b = bytes.TrimSpace(b)
reader := bufio.NewReader(bytes.NewBuffer(b)) openingLine := parser.Opening()
closingLine := parser.Closing()
// Read first line, which should indicate metadata or not if !bytes.HasPrefix(b, openingLine) {
line, err := reader.ReadBytes('\n')
if err != nil || !bytes.Equal(bytes.TrimSpace(line), parser.Opening()) {
return nil, b, fmt.Errorf("first line missing expected metadata identifier") return nil, b, fmt.Errorf("first line missing expected metadata identifier")
} }
metaStart := len(openingLine)
// buffer for metadata contents if _, ok := parser.(*JSONMetadataParser); ok {
metaBuf := bytes.Buffer{} metaStart = 0
// Read remaining lines until closing identifier is found
for {
line, err := reader.ReadBytes('\n')
if err != nil && err != io.EOF {
return nil, nil, err
}
// if closing identifier found, the remaining bytes must be markdown content
if bytes.Equal(bytes.TrimSpace(line), parser.Closing()) {
break
} }
metaEnd := bytes.Index(b[metaStart:], closingLine)
// if file ended, by this point no closing identifier was found if metaEnd == -1 {
if err == io.EOF {
return nil, nil, fmt.Errorf("metadata not closed ('%s' not found)", parser.Closing()) return nil, nil, fmt.Errorf("metadata not closed ('%s' not found)", parser.Closing())
} }
metaEnd += metaStart
metaBuf.Write(line) if _, ok := parser.(*JSONMetadataParser); ok {
metaBuf.WriteString("\r\n") metaEnd += len(closingLine)
} }
metadata = b[metaStart:metaEnd]
// By now, the rest of the buffer contains markdown content markdown = b[metaEnd:]
contentBuf := new(bytes.Buffer) if _, ok := parser.(*JSONMetadataParser); !ok {
io.Copy(contentBuf, reader) markdown = b[metaEnd+len(closingLine):]
}
return metaBuf.Bytes(), contentBuf.Bytes(), nil return metadata, markdown, nil
} }
// findParser finds the parser using line that contains opening identifier // findParser finds the parser using line that contains opening identifier
......
...@@ -8,7 +8,13 @@ import ( ...@@ -8,7 +8,13 @@ import (
"testing" "testing"
) )
var TOML = [4]string{` func check(t *testing.T, err error) {
if err != nil {
t.Fatal(err)
}
}
var TOML = [5]string{`
title = "A title" title = "A title"
template = "default" template = "default"
name = "value" name = "value"
...@@ -19,16 +25,22 @@ template = "default" ...@@ -19,16 +25,22 @@ template = "default"
name = "value" name = "value"
+++ +++
Page content Page content
`, `,
`+++ `+++
title = "A title" title = "A title"
template = "default" template = "default"
name = "value" name = "value"
`, `,
`title = "A title" template = "default" [variables] name = "value"`, `title = "A title" template = "default" [variables] name = "value"`,
`+++
title = "A title"
template = "default"
name = "value"
+++
`,
} }
var YAML = [4]string{` var YAML = [5]string{`
title : A title title : A title
template : default template : default
name : value name : value
...@@ -38,16 +50,23 @@ title : A title ...@@ -38,16 +50,23 @@ title : A title
template : default template : default
name : value name : value
--- ---
Page content Page content
`, `,
`--- `---
title : A title title : A title
template : default template : default
name : value name : value
`, `,
`title : A title template : default variables : name : value`, `title : A title template : default variables : name : value`,
`---
title : A title
template : default
name : value
---
`,
} }
var JSON = [4]string{`
var JSON = [5]string{`
"title" : "A title", "title" : "A title",
"template" : "default", "template" : "default",
"name" : "value" "name" : "value"
...@@ -58,26 +77,26 @@ var JSON = [4]string{` ...@@ -58,26 +77,26 @@ var JSON = [4]string{`
"name" : "value" "name" : "value"
} }
Page content Page content
`, `,
` `
{ {
"title" : "A title", "title" : "A title",
"template" : "default", "template" : "default",
"name" : "value" "name" : "value"
`, `,
` `
{{ {
"title" : "A title", "title" :: "A title",
"template" : "default", "template" : "default",
"name" : "value" "name" : "value"
} }
`, `,
`{
"title" : "A title",
"template" : "default",
"name" : "value"
} }
`,
func check(t *testing.T, err error) {
if err != nil {
t.Fatal(err)
}
} }
func TestParsers(t *testing.T) { func TestParsers(t *testing.T) {
...@@ -107,7 +126,7 @@ func TestParsers(t *testing.T) { ...@@ -107,7 +126,7 @@ func TestParsers(t *testing.T) {
data := []struct { data := []struct {
parser MetadataParser parser MetadataParser
testData [4]string testData [5]string
name string name string
}{ }{
{&JSONMetadataParser{metadata: Metadata{Variables: make(map[string]string)}}, JSON, "json"}, {&JSONMetadataParser{metadata: Metadata{Variables: make(map[string]string)}}, JSON, "json"},
...@@ -150,6 +169,52 @@ func TestParsers(t *testing.T) { ...@@ -150,6 +169,52 @@ func TestParsers(t *testing.T) {
if md, err = v.parser.Parse([]byte(v.testData[3])); err == nil { if md, err = v.parser.Parse([]byte(v.testData[3])); err == nil {
t.Fatalf("Expected error for invalid metadata for %v", v.name) t.Fatalf("Expected error for invalid metadata for %v", v.name)
} }
// front matter but no body
if md, err = v.parser.Parse([]byte(v.testData[4])); err != nil {
t.Fatalf("Unexpected error for valid metadata but no body for %v", v.name)
}
} }
} }
func TestLargeBody(t *testing.T) {
var JSON = `{
"template": "chapter"
}
Mycket olika byggnader har man i de nordiska rikena: pyramidformiga, kilformiga, välvda, runda och fyrkantiga. De pyramidformiga består helt enkelt av träribbor, som upptill löper samman och nedtill bildar en vidare krets; de är avsedda att användas av hantverkarna under sommaren, för att de inte ska plågas av solen, på samma gång som de besväras av rök och eld. De kilformiga husen är i regel försedda med höga tak, för att de täta och tunga snömassorna fortare ska kunna blåsa av och inte tynga ned taken. Dessa är täckta av björknäver, tegel eller kluvet spån av furu - för kådans skull -, gran, ek eller bok; taken på de förmögnas hus däremot med plåtar av koppar eller bly, i likhet med kyrktaken. Valvbyggnaderna uppförs ganska konstnärligt till skydd mot våldsamma vindar och snöfall, görs av sten eller trä, och är avsedda för olika alldagliga viktiga ändamål. Liknande byggnader kan finnas i stormännens gårdar där de används som förvaringsrum för husgeråd och jordbruksredskap. De runda byggnaderna - som för övrigt är de högst sällsynta - används av konstnärer, som vid sitt arbete behöver ett jämnt fördelat ljus från taket. Vanligast är de fyrkantiga husen, vars grova bjälkar är synnerligen väl hopfogade i hörnen - ett sant mästerverk av byggnadskonst; även dessa har fönster högt uppe i taken, för att dagsljuset skall kunna strömma in och ge alla därinne full belysning. Stenhusen har dörröppningar i förhållande till byggnadens storlek, men smala fönstergluggar, som skydd mot den stränga kölden, frosten och snön. Vore de större och vidare, såsom fönstren i Italien, skulle husen i följd av den fint yrande snön, som röres upp av den starka blåsten, precis som dammet av virvelvinden, snart nog fyllas med massor av snö och inte kunna stå emot dess tryck, utan störta samman.
`
var TOML = `+++
template = "chapter"
+++
Mycket olika byggnader har man i de nordiska rikena: pyramidformiga, kilformiga, välvda, runda och fyrkantiga. De pyramidformiga består helt enkelt av träribbor, som upptill löper samman och nedtill bildar en vidare krets; de är avsedda att användas av hantverkarna under sommaren, för att de inte ska plågas av solen, på samma gång som de besväras av rök och eld. De kilformiga husen är i regel försedda med höga tak, för att de täta och tunga snömassorna fortare ska kunna blåsa av och inte tynga ned taken. Dessa är täckta av björknäver, tegel eller kluvet spån av furu - för kådans skull -, gran, ek eller bok; taken på de förmögnas hus däremot med plåtar av koppar eller bly, i likhet med kyrktaken. Valvbyggnaderna uppförs ganska konstnärligt till skydd mot våldsamma vindar och snöfall, görs av sten eller trä, och är avsedda för olika alldagliga viktiga ändamål. Liknande byggnader kan finnas i stormännens gårdar där de används som förvaringsrum för husgeråd och jordbruksredskap. De runda byggnaderna - som för övrigt är de högst sällsynta - används av konstnärer, som vid sitt arbete behöver ett jämnt fördelat ljus från taket. Vanligast är de fyrkantiga husen, vars grova bjälkar är synnerligen väl hopfogade i hörnen - ett sant mästerverk av byggnadskonst; även dessa har fönster högt uppe i taken, för att dagsljuset skall kunna strömma in och ge alla därinne full belysning. Stenhusen har dörröppningar i förhållande till byggnadens storlek, men smala fönstergluggar, som skydd mot den stränga kölden, frosten och snön. Vore de större och vidare, såsom fönstren i Italien, skulle husen i följd av den fint yrande snön, som röres upp av den starka blåsten, precis som dammet av virvelvinden, snart nog fyllas med massor av snö och inte kunna stå emot dess tryck, utan störta samman.
`
var YAML = `---
template : chapter
---
Mycket olika byggnader har man i de nordiska rikena: pyramidformiga, kilformiga, välvda, runda och fyrkantiga. De pyramidformiga består helt enkelt av träribbor, som upptill löper samman och nedtill bildar en vidare krets; de är avsedda att användas av hantverkarna under sommaren, för att de inte ska plågas av solen, på samma gång som de besväras av rök och eld. De kilformiga husen är i regel försedda med höga tak, för att de täta och tunga snömassorna fortare ska kunna blåsa av och inte tynga ned taken. Dessa är täckta av björknäver, tegel eller kluvet spån av furu - för kådans skull -, gran, ek eller bok; taken på de förmögnas hus däremot med plåtar av koppar eller bly, i likhet med kyrktaken. Valvbyggnaderna uppförs ganska konstnärligt till skydd mot våldsamma vindar och snöfall, görs av sten eller trä, och är avsedda för olika alldagliga viktiga ändamål. Liknande byggnader kan finnas i stormännens gårdar där de används som förvaringsrum för husgeråd och jordbruksredskap. De runda byggnaderna - som för övrigt är de högst sällsynta - används av konstnärer, som vid sitt arbete behöver ett jämnt fördelat ljus från taket. Vanligast är de fyrkantiga husen, vars grova bjälkar är synnerligen väl hopfogade i hörnen - ett sant mästerverk av byggnadskonst; även dessa har fönster högt uppe i taken, för att dagsljuset skall kunna strömma in och ge alla därinne full belysning. Stenhusen har dörröppningar i förhållande till byggnadens storlek, men smala fönstergluggar, som skydd mot den stränga kölden, frosten och snön. Vore de större och vidare, såsom fönstren i Italien, skulle husen i följd av den fint yrande snön, som röres upp av den starka blåsten, precis som dammet av virvelvinden, snart nog fyllas med massor av snö och inte kunna stå emot dess tryck, utan störta samman.
`
var expectedBody = `Mycket olika byggnader har man i de nordiska rikena: pyramidformiga, kilformiga, välvda, runda och fyrkantiga. De pyramidformiga består helt enkelt av träribbor, som upptill löper samman och nedtill bildar en vidare krets; de är avsedda att användas av hantverkarna under sommaren, för att de inte ska plågas av solen, på samma gång som de besväras av rök och eld. De kilformiga husen är i regel försedda med höga tak, för att de täta och tunga snömassorna fortare ska kunna blåsa av och inte tynga ned taken. Dessa är täckta av björknäver, tegel eller kluvet spån av furu - för kådans skull -, gran, ek eller bok; taken på de förmögnas hus däremot med plåtar av koppar eller bly, i likhet med kyrktaken. Valvbyggnaderna uppförs ganska konstnärligt till skydd mot våldsamma vindar och snöfall, görs av sten eller trä, och är avsedda för olika alldagliga viktiga ändamål. Liknande byggnader kan finnas i stormännens gårdar där de används som förvaringsrum för husgeråd och jordbruksredskap. De runda byggnaderna - som för övrigt är de högst sällsynta - används av konstnärer, som vid sitt arbete behöver ett jämnt fördelat ljus från taket. Vanligast är de fyrkantiga husen, vars grova bjälkar är synnerligen väl hopfogade i hörnen - ett sant mästerverk av byggnadskonst; även dessa har fönster högt uppe i taken, för att dagsljuset skall kunna strömma in och ge alla därinne full belysning. Stenhusen har dörröppningar i förhållande till byggnadens storlek, men smala fönstergluggar, som skydd mot den stränga kölden, frosten och snön. Vore de större och vidare, såsom fönstren i Italien, skulle husen i följd av den fint yrande snön, som röres upp av den starka blåsten, precis som dammet av virvelvinden, snart nog fyllas med massor av snö och inte kunna stå emot dess tryck, utan störta samman.
`
data := []struct {
parser MetadataParser
testData string
name string
}{
{&JSONMetadataParser{metadata: Metadata{Variables: make(map[string]string)}}, JSON, "json"},
{&YAMLMetadataParser{metadata: Metadata{Variables: make(map[string]string)}}, YAML, "yaml"},
{&TOMLMetadataParser{metadata: Metadata{Variables: make(map[string]string)}}, TOML, "toml"},
}
for _, v := range data {
// metadata without identifiers
if md, err := v.parser.Parse([]byte(v.testData)); err != nil || strings.TrimSpace(string(md)) != strings.TrimSpace(expectedBody) {
t.Fatalf("Error not expected and/or markdown not equal for %v", v.name)
}
}
}
...@@ -13,14 +13,14 @@ func TestWatcher(t *testing.T) { ...@@ -13,14 +13,14 @@ func TestWatcher(t *testing.T) {
interval := time.Millisecond * 100 interval := time.Millisecond * 100
i := 0 i := 0
out := "" out := ""
syncChan := make(chan struct{})
stopChan := TickerFunc(interval, func() { stopChan := TickerFunc(interval, func() {
i++ i++
out += fmt.Sprint(i) out += fmt.Sprint(i)
syncChan <- struct{}{}
}) })
// wait little more because of concurrency sleepInSync(8, syncChan, stopChan)
time.Sleep(interval * 9) if out != expected {
stopChan <- struct{}{}
if !strings.HasPrefix(out, expected) {
t.Fatalf("Expected to have prefix %v, found %v", expected, out) t.Fatalf("Expected to have prefix %v, found %v", expected, out)
} }
out = "" out = ""
...@@ -31,8 +31,9 @@ func TestWatcher(t *testing.T) { ...@@ -31,8 +31,9 @@ func TestWatcher(t *testing.T) {
mu.Lock() mu.Lock()
out += fmt.Sprint(i) out += fmt.Sprint(i)
mu.Unlock() mu.Unlock()
syncChan <- struct{}{}
}) })
time.Sleep(interval * 10) sleepInSync(9, syncChan, stopChan)
mu.Lock() mu.Lock()
res := out res := out
mu.Unlock() mu.Unlock()
...@@ -40,3 +41,10 @@ func TestWatcher(t *testing.T) { ...@@ -40,3 +41,10 @@ func TestWatcher(t *testing.T) {
t.Fatalf("expected (%v) must be a proper prefix of out(%v).", expected, out) t.Fatalf("expected (%v) must be a proper prefix of out(%v).", expected, out)
} }
} }
func sleepInSync(times int, syncChan chan struct{}, stopChan chan struct{}) {
for i := 0; i < times; i++ {
<-syncChan
}
stopChan <- struct{}{}
}
...@@ -3,6 +3,7 @@ package middleware ...@@ -3,6 +3,7 @@ package middleware
import ( import (
"net" "net"
"net/http" "net/http"
"net/url"
"path" "path"
"strconv" "strconv"
"strings" "strings"
...@@ -15,6 +16,7 @@ import ( ...@@ -15,6 +16,7 @@ import (
// NewReplacer to get one of these. // NewReplacer to get one of these.
type Replacer interface { type Replacer interface {
Replace(string) string Replace(string) string
Set(key, value string)
} }
type replacer struct { type replacer struct {
...@@ -40,7 +42,9 @@ func NewReplacer(r *http.Request, rr *responseRecorder, emptyValue string) Repla ...@@ -40,7 +42,9 @@ func NewReplacer(r *http.Request, rr *responseRecorder, emptyValue string) Repla
}(), }(),
"{host}": r.Host, "{host}": r.Host,
"{path}": r.URL.Path, "{path}": r.URL.Path,
"{path_escaped}": url.QueryEscape(r.URL.Path),
"{query}": r.URL.RawQuery, "{query}": r.URL.RawQuery,
"{query_escaped}": url.QueryEscape(r.URL.RawQuery),
"{fragment}": r.URL.Fragment, "{fragment}": r.URL.Fragment,
"{proto}": r.Proto, "{proto}": r.Proto,
"{remote}": func() string { "{remote}": func() string {
...@@ -61,6 +65,7 @@ func NewReplacer(r *http.Request, rr *responseRecorder, emptyValue string) Repla ...@@ -61,6 +65,7 @@ func NewReplacer(r *http.Request, rr *responseRecorder, emptyValue string) Repla
return port return port
}(), }(),
"{uri}": r.URL.RequestURI(), "{uri}": r.URL.RequestURI(),
"{uri_escaped}": url.QueryEscape(r.URL.RequestURI()),
"{when}": func() string { "{when}": func() string {
return time.Now().Format(timeFormat) return time.Now().Format(timeFormat)
}(), }(),
...@@ -81,9 +86,9 @@ func NewReplacer(r *http.Request, rr *responseRecorder, emptyValue string) Repla ...@@ -81,9 +86,9 @@ func NewReplacer(r *http.Request, rr *responseRecorder, emptyValue string) Repla
rep.replacements["{latency}"] = time.Since(rr.start).String() rep.replacements["{latency}"] = time.Since(rr.start).String()
} }
// Header placeholders // Header placeholders (case-insensitive)
for header, val := range r.Header { for header, values := range r.Header {
rep.replacements[headerReplacer+header+"}"] = strings.Join(val, ",") rep.replacements[headerReplacer+strings.ToLower(header)+"}"] = strings.Join(values, ",")
} }
return rep return rep
...@@ -92,27 +97,39 @@ func NewReplacer(r *http.Request, rr *responseRecorder, emptyValue string) Repla ...@@ -92,27 +97,39 @@ func NewReplacer(r *http.Request, rr *responseRecorder, emptyValue string) Repla
// Replace performs a replacement of values on s and returns // Replace performs a replacement of values on s and returns
// the string with the replaced values. // the string with the replaced values.
func (r replacer) Replace(s string) string { func (r replacer) Replace(s string) string {
for placeholder, replacement := range r.replacements { // Header replacements - these are case-insensitive, so we can't just use strings.Replace()
if replacement == "" {
replacement = r.emptyValue
}
s = strings.Replace(s, placeholder, replacement, -1)
}
// Replace any header placeholders that weren't found
for strings.Contains(s, headerReplacer) { for strings.Contains(s, headerReplacer) {
idxStart := strings.Index(s, headerReplacer) idxStart := strings.Index(s, headerReplacer)
endOffset := idxStart + len(headerReplacer) endOffset := idxStart + len(headerReplacer)
idxEnd := strings.Index(s[endOffset:], "}") idxEnd := strings.Index(s[endOffset:], "}")
if idxEnd > -1 { if idxEnd > -1 {
s = s[:idxStart] + r.emptyValue + s[endOffset+idxEnd+1:] placeholder := strings.ToLower(s[idxStart : endOffset+idxEnd+1])
replacement := r.replacements[placeholder]
if replacement == "" {
replacement = r.emptyValue
}
s = s[:idxStart] + replacement + s[endOffset+idxEnd+1:]
} else { } else {
break break
} }
} }
// Regular replacements - these are easier because they're case-sensitive
for placeholder, replacement := range r.replacements {
if replacement == "" {
replacement = r.emptyValue
}
s = strings.Replace(s, placeholder, replacement, -1)
}
return s return s
} }
// Set sets key to value in the replacements map.
func (r replacer) Set(key, value string) {
r.replacements["{"+key+"}"] = value
}
const ( const (
timeFormat = "02/Jan/2006:15:04:05 -0700" timeFormat = "02/Jan/2006:15:04:05 -0700"
headerReplacer = "{>" headerReplacer = "{>"
......
...@@ -10,62 +10,115 @@ import ( ...@@ -10,62 +10,115 @@ import (
func TestNewReplacer(t *testing.T) { func TestNewReplacer(t *testing.T) {
w := httptest.NewRecorder() w := httptest.NewRecorder()
recordRequest := NewResponseRecorder(w) recordRequest := NewResponseRecorder(w)
userJSON := `{"username": "dennis"}` reader := strings.NewReader(`{"username": "dennis"}`)
reader := strings.NewReader(userJSON) //Convert string to reader request, err := http.NewRequest("POST", "http://localhost", reader)
request, err := http.NewRequest("POST", "http://caddyserver.com", reader) //Create request with JSON body
if err != nil { if err != nil {
t.Fatalf("Request Formation Failed \n") t.Fatal("Request Formation Failed\n")
} }
replaceValues := NewReplacer(request, recordRequest, "") replaceValues := NewReplacer(request, recordRequest, "")
switch v := replaceValues.(type) { switch v := replaceValues.(type) {
case replacer: case replacer:
if v.replacements["{host}"] != "caddyserver.com" { if v.replacements["{host}"] != "localhost" {
t.Errorf("Expected host to be caddyserver.com") t.Error("Expected host to be localhost")
} }
if v.replacements["{method}"] != "POST" { if v.replacements["{method}"] != "POST" {
t.Errorf("Expected request method to be POST") t.Error("Expected request method to be POST")
} }
if v.replacements["{status}"] != "200" { if v.replacements["{status}"] != "200" {
t.Errorf("Expected status to be 200") t.Error("Expected status to be 200")
} }
default: default:
t.Fatalf("Return Value from New Replacer expected pass type assertion into a replacer type \n") t.Fatal("Return Value from New Replacer expected pass type assertion into a replacer type\n")
} }
} }
func TestReplace(t *testing.T) { func TestReplace(t *testing.T) {
w := httptest.NewRecorder() w := httptest.NewRecorder()
recordRequest := NewResponseRecorder(w) recordRequest := NewResponseRecorder(w)
userJSON := `{"username": "dennis"}` reader := strings.NewReader(`{"username": "dennis"}`)
reader := strings.NewReader(userJSON) //Convert string to reader request, err := http.NewRequest("POST", "http://localhost", reader)
request, err := http.NewRequest("POST", "http://caddyserver.com", reader) //Create request with JSON body
if err != nil { if err != nil {
t.Fatalf("Request Formation Failed \n") t.Fatal("Request Formation Failed\n")
} }
replaceValues := NewReplacer(request, recordRequest, "") request.Header.Set("Custom", "foobarbaz")
request.Header.Set("ShorterVal", "1")
repl := NewReplacer(request, recordRequest, "-")
switch v := replaceValues.(type) { if expected, actual := "This host is localhost.", repl.Replace("This host is {host}."); expected != actual {
case replacer: t.Errorf("{host} replacement: expected '%s', got '%s'", expected, actual)
}
if expected, actual := "This request method is POST.", repl.Replace("This request method is {method}."); expected != actual {
t.Errorf("{method} replacement: expected '%s', got '%s'", expected, actual)
}
if expected, actual := "The response status is 200.", repl.Replace("The response status is {status}."); expected != actual {
t.Errorf("{status} replacement: expected '%s', got '%s'", expected, actual)
}
if expected, actual := "The Custom header is foobarbaz.", repl.Replace("The Custom header is {>Custom}."); expected != actual {
t.Errorf("{>Custom} replacement: expected '%s', got '%s'", expected, actual)
}
// Test header case-insensitivity
if expected, actual := "The cUsToM header is foobarbaz...", repl.Replace("The cUsToM header is {>cUsToM}..."); expected != actual {
t.Errorf("{>cUsToM} replacement: expected '%s', got '%s'", expected, actual)
}
if v.Replace("This host is {host}") != "This host is caddyserver.com" { // Test non-existent header/value
t.Errorf("Expected host replacement failed") if expected, actual := "The Non-Existent header is -.", repl.Replace("The Non-Existent header is {>Non-Existent}."); expected != actual {
t.Errorf("{>Non-Existent} replacement: expected '%s', got '%s'", expected, actual)
} }
if v.Replace("This request method is {method}") != "This request method is POST" {
t.Errorf("Expected method replacement failed") // Test bad placeholder
if expected, actual := "Bad {host placeholder...", repl.Replace("Bad {host placeholder..."); expected != actual {
t.Errorf("bad placeholder: expected '%s', got '%s'", expected, actual)
} }
if v.Replace("The response status is {status}") != "The response status is 200" {
t.Errorf("Expected status replacement failed") // Test bad header placeholder
if expected, actual := "Bad {>Custom placeholder", repl.Replace("Bad {>Custom placeholder"); expected != actual {
t.Errorf("bad header placeholder: expected '%s', got '%s'", expected, actual)
} }
default: // Test bad header placeholder with valid one later
t.Fatalf("Return Value from New Replacer expected pass type assertion into a replacer type \n") if expected, actual := "Bad -", repl.Replace("Bad {>Custom placeholder {>ShorterVal}"); expected != actual {
t.Errorf("bad header placeholders: expected '%s', got '%s'", expected, actual)
} }
// Test shorter header value with multiple placeholders
if expected, actual := "Short value 1 then foobarbaz.", repl.Replace("Short value {>ShorterVal} then {>Custom}."); expected != actual {
t.Errorf("short value: expected '%s', got '%s'", expected, actual)
}
}
func TestSet(t *testing.T) {
w := httptest.NewRecorder()
recordRequest := NewResponseRecorder(w)
reader := strings.NewReader(`{"username": "dennis"}`)
request, err := http.NewRequest("POST", "http://localhost", reader)
if err != nil {
t.Fatalf("Request Formation Failed \n")
}
repl := NewReplacer(request, recordRequest, "")
repl.Set("host", "getcaddy.com")
repl.Set("method", "GET")
repl.Set("status", "201")
repl.Set("variable", "value")
if repl.Replace("This host is {host}") != "This host is getcaddy.com" {
t.Error("Expected host replacement failed")
}
if repl.Replace("This request method is {method}") != "This request method is GET" {
t.Error("Expected method replacement failed")
}
if repl.Replace("The response status is {status}") != "The response status is 201" {
t.Error("Expected status replacement failed")
}
if repl.Replace("The value of variable is {variable}") != "The value of variable is value" {
t.Error("Expected variable replacement failed")
}
} }
package rewrite
import (
"fmt"
"net/http"
"regexp"
"strings"
"github.com/mholt/caddy/middleware"
)
const (
// Operators
Is = "is"
Not = "not"
Has = "has"
StartsWith = "starts_with"
EndsWith = "ends_with"
Match = "match"
)
func operatorError(operator string) error {
return fmt.Errorf("Invalid operator %v", operator)
}
func newReplacer(r *http.Request) middleware.Replacer {
return middleware.NewReplacer(r, nil, "")
}
// condition is a rewrite condition.
type condition func(string, string) bool
var conditions = map[string]condition{
Is: isFunc,
Not: notFunc,
Has: hasFunc,
StartsWith: startsWithFunc,
EndsWith: endsWithFunc,
Match: matchFunc,
}
// isFunc is condition for Is operator.
// It checks for equality.
func isFunc(a, b string) bool {
return a == b
}
// notFunc is condition for Not operator.
// It checks for inequality.
func notFunc(a, b string) bool {
return a != b
}
// hasFunc is condition for Has operator.
// It checks if b is a substring of a.
func hasFunc(a, b string) bool {
return strings.Contains(a, b)
}
// startsWithFunc is condition for StartsWith operator.
// It checks if b is a prefix of a.
func startsWithFunc(a, b string) bool {
return strings.HasPrefix(a, b)
}
// endsWithFunc is condition for EndsWith operator.
// It checks if b is a suffix of a.
func endsWithFunc(a, b string) bool {
return strings.HasSuffix(a, b)
}
// matchFunc is condition for Match operator.
// It does regexp matching of a against pattern in b
func matchFunc(a, b string) bool {
matched, _ := regexp.MatchString(b, a)
return matched
}
// If is statement for a rewrite condition.
type If struct {
A string
Operator string
B string
}
// True returns true if the condition is true and false otherwise.
// If r is not nil, it replaces placeholders before comparison.
func (i If) True(r *http.Request) bool {
if c, ok := conditions[i.Operator]; ok {
a, b := i.A, i.B
if r != nil {
replacer := newReplacer(r)
a = replacer.Replace(i.A)
b = replacer.Replace(i.B)
}
return c(a, b)
}
return false
}
// NewIf creates a new If condition.
func NewIf(a, operator, b string) (If, error) {
if _, ok := conditions[operator]; !ok {
return If{}, operatorError(operator)
}
return If{
A: a,
Operator: operator,
B: b,
}, nil
}
package rewrite
import (
"net/http"
"strings"
"testing"
)
func TestConditions(t *testing.T) {
tests := []struct {
condition string
isTrue bool
}{
{"a is b", false},
{"a is a", true},
{"a not b", true},
{"a not a", false},
{"a has a", true},
{"a has b", false},
{"ba has b", true},
{"bab has b", true},
{"bab has bb", false},
{"bab starts_with bb", false},
{"bab starts_with ba", true},
{"bab starts_with bab", true},
{"bab ends_with bb", false},
{"bab ends_with bab", true},
{"bab ends_with ab", true},
{"a match *", false},
{"a match a", true},
{"a match .*", true},
{"a match a.*", true},
{"a match b.*", false},
{"ba match b.*", true},
{"ba match b[a-z]", true},
{"b0 match b[a-z]", false},
{"b0a match b[a-z]", false},
{"b0a match b[a-z]+", false},
{"b0a match b[a-z0-9]+", true},
}
for i, test := range tests {
str := strings.Fields(test.condition)
ifCond, err := NewIf(str[0], str[1], str[2])
if err != nil {
t.Error(err)
}
isTrue := ifCond.True(nil)
if isTrue != test.isTrue {
t.Errorf("Test %v: expected %v found %v", i, test.isTrue, isTrue)
}
}
invalidOperators := []string{"ss", "and", "if"}
for _, op := range invalidOperators {
_, err := NewIf("a", op, "b")
if err == nil {
t.Errorf("Invalid operator %v used, expected error.", op)
}
}
replaceTests := []struct {
url string
condition string
isTrue bool
}{
{"/home", "{uri} match /home", true},
{"/hom", "{uri} match /home", false},
{"/hom", "{uri} starts_with /home", false},
{"/hom", "{uri} starts_with /h", true},
{"/home/.hiddenfile", `{uri} match \/\.(.*)`, true},
{"/home/.hiddendir/afile", `{uri} match \/\.(.*)`, true},
}
for i, test := range replaceTests {
r, err := http.NewRequest("GET", test.url, nil)
if err != nil {
t.Error(err)
}
str := strings.Fields(test.condition)
ifCond, err := NewIf(str[0], str[1], str[2])
if err != nil {
t.Error(err)
}
isTrue := ifCond.True(r)
if isTrue != test.isTrue {
t.Errorf("Test %v: expected %v found %v", i, test.isTrue, isTrue)
}
}
}
...@@ -5,7 +5,6 @@ package rewrite ...@@ -5,7 +5,6 @@ package rewrite
import ( import (
"fmt" "fmt"
"net/http" "net/http"
"net/url"
"path" "path"
"path/filepath" "path/filepath"
"regexp" "regexp"
...@@ -17,13 +16,14 @@ import ( ...@@ -17,13 +16,14 @@ import (
// Rewrite is middleware to rewrite request locations internally before being handled. // Rewrite is middleware to rewrite request locations internally before being handled.
type Rewrite struct { type Rewrite struct {
Next middleware.Handler Next middleware.Handler
FileSys http.FileSystem
Rules []Rule Rules []Rule
} }
// ServeHTTP implements the middleware.Handler interface. // ServeHTTP implements the middleware.Handler interface.
func (rw Rewrite) ServeHTTP(w http.ResponseWriter, r *http.Request) (int, error) { func (rw Rewrite) ServeHTTP(w http.ResponseWriter, r *http.Request) (int, error) {
for _, rule := range rw.Rules { for _, rule := range rw.Rules {
if ok := rule.Rewrite(r); ok { if ok := rule.Rewrite(rw.FileSys, r); ok {
break break
} }
} }
...@@ -33,7 +33,7 @@ func (rw Rewrite) ServeHTTP(w http.ResponseWriter, r *http.Request) (int, error) ...@@ -33,7 +33,7 @@ func (rw Rewrite) ServeHTTP(w http.ResponseWriter, r *http.Request) (int, error)
// Rule describes an internal location rewrite rule. // Rule describes an internal location rewrite rule.
type Rule interface { type Rule interface {
// Rewrite rewrites the internal location of the current request. // Rewrite rewrites the internal location of the current request.
Rewrite(*http.Request) bool Rewrite(http.FileSystem, *http.Request) bool
} }
// SimpleRule is a simple rewrite rule. // SimpleRule is a simple rewrite rule.
...@@ -47,23 +47,20 @@ func NewSimpleRule(from, to string) SimpleRule { ...@@ -47,23 +47,20 @@ func NewSimpleRule(from, to string) SimpleRule {
} }
// Rewrite rewrites the internal location of the current request. // Rewrite rewrites the internal location of the current request.
func (s SimpleRule) Rewrite(r *http.Request) bool { func (s SimpleRule) Rewrite(fs http.FileSystem, r *http.Request) bool {
if s.From == r.URL.Path { if s.From == r.URL.Path {
// take note of this rewrite for internal use by fastcgi // take note of this rewrite for internal use by fastcgi
// all we need is the URI, not full URL // all we need is the URI, not full URL
r.Header.Set(headerFieldName, r.URL.RequestURI()) r.Header.Set(headerFieldName, r.URL.RequestURI())
// replace variables // attempt rewrite
to := path.Clean(middleware.NewReplacer(r, nil, "").Replace(s.To)) return To(fs, r, s.To, newReplacer(r))
r.URL.Path = to
return true
} }
return false return false
} }
// RegexpRule is a rewrite rule based on a regular expression // ComplexRule is a rewrite rule based on a regular expression
type RegexpRule struct { type ComplexRule struct {
// Path base. Request to this path and subpaths will be rewritten // Path base. Request to this path and subpaths will be rewritten
Base string Base string
...@@ -73,18 +70,26 @@ type RegexpRule struct { ...@@ -73,18 +70,26 @@ type RegexpRule struct {
// Extensions to filter by // Extensions to filter by
Exts []string Exts []string
// Rewrite conditions
Ifs []If
*regexp.Regexp *regexp.Regexp
} }
// NewRegexpRule creates a new RegexpRule. It returns an error if regexp // NewRegexpRule creates a new RegexpRule. It returns an error if regexp
// pattern (pattern) or extensions (ext) are invalid. // pattern (pattern) or extensions (ext) are invalid.
func NewRegexpRule(base, pattern, to string, ext []string) (*RegexpRule, error) { func NewComplexRule(base, pattern, to string, ext []string, ifs []If) (*ComplexRule, error) {
r, err := regexp.Compile(pattern) // validate regexp if present
var r *regexp.Regexp
if pattern != "" {
var err error
r, err = regexp.Compile(pattern)
if err != nil { if err != nil {
return nil, err return nil, err
} }
}
// validate extensions // validate extensions if present
for _, v := range ext { for _, v := range ext {
if len(v) < 2 || (len(v) < 3 && v[0] == '!') { if len(v) < 2 || (len(v) < 3 && v[0] == '!') {
// check if no extension is specified // check if no extension is specified
...@@ -94,17 +99,19 @@ func NewRegexpRule(base, pattern, to string, ext []string) (*RegexpRule, error) ...@@ -94,17 +99,19 @@ func NewRegexpRule(base, pattern, to string, ext []string) (*RegexpRule, error)
} }
} }
return &RegexpRule{ return &ComplexRule{
base, Base: base,
to, To: to,
ext, Exts: ext,
r, Ifs: ifs,
Regexp: r,
}, nil }, nil
} }
// Rewrite rewrites the internal location of the current request. // Rewrite rewrites the internal location of the current request.
func (r *RegexpRule) Rewrite(req *http.Request) bool { func (r *ComplexRule) Rewrite(fs http.FileSystem, req *http.Request) bool {
rPath := req.URL.Path rPath := req.URL.Path
replacer := newReplacer(req)
// validate base // validate base
if !middleware.Path(rPath).Matches(r.Base) { if !middleware.Path(rPath).Matches(r.Base) {
...@@ -122,36 +129,35 @@ func (r *RegexpRule) Rewrite(req *http.Request) bool { ...@@ -122,36 +129,35 @@ func (r *RegexpRule) Rewrite(req *http.Request) bool {
start-- start--
} }
// validate regexp // validate regexp if present
if !r.MatchString(rPath[start:]) { if r.Regexp != nil {
matches := r.FindStringSubmatch(rPath[start:])
switch len(matches) {
case 0:
// no match
return false return false
default:
// set regexp match variables {1}, {2} ...
for i := 1; i < len(matches); i++ {
replacer.Set(fmt.Sprint(i), matches[i])
}
}
} }
// replace variables // validate rewrite conditions
to := path.Clean(middleware.NewReplacer(req, nil, "").Replace(r.To)) for _, i := range r.Ifs {
if !i.True(req) {
// validate resulting path
url, err := url.Parse(to)
if err != nil {
return false return false
} }
// take note of this rewrite for internal use by fastcgi
// all we need is the URI, not full URL
req.Header.Set(headerFieldName, req.URL.RequestURI())
// perform rewrite
req.URL.Path = url.Path
if url.RawQuery != "" {
// overwrite query string if present
req.URL.RawQuery = url.RawQuery
} }
return true
// attempt rewrite
return To(fs, req, r.To, replacer)
} }
// matchExt matches rPath against registered file extensions. // matchExt matches rPath against registered file extensions.
// Returns true if a match is found and false otherwise. // Returns true if a match is found and false otherwise.
func (r *RegexpRule) matchExt(rPath string) bool { func (r *ComplexRule) matchExt(rPath string) bool {
f := filepath.Base(rPath) f := filepath.Base(rPath)
ext := path.Ext(f) ext := path.Ext(f)
if ext == "" { if ext == "" {
......
...@@ -4,9 +4,8 @@ import ( ...@@ -4,9 +4,8 @@ import (
"fmt" "fmt"
"net/http" "net/http"
"net/http/httptest" "net/http/httptest"
"testing"
"strings" "strings"
"testing"
"github.com/mholt/caddy/middleware" "github.com/mholt/caddy/middleware"
) )
...@@ -19,9 +18,10 @@ func TestRewrite(t *testing.T) { ...@@ -19,9 +18,10 @@ func TestRewrite(t *testing.T) {
NewSimpleRule("/a", "/b"), NewSimpleRule("/a", "/b"),
NewSimpleRule("/b", "/b{uri}"), NewSimpleRule("/b", "/b{uri}"),
}, },
FileSys: http.Dir("."),
} }
regexpRules := [][]string{ regexps := [][]string{
{"/reg/", ".*", "/to", ""}, {"/reg/", ".*", "/to", ""},
{"/r/", "[a-z]+", "/toaz", "!.html|"}, {"/r/", "[a-z]+", "/toaz", "!.html|"},
{"/url/", "a([a-z0-9]*)s([A-Z]{2})", "/to/{path}", ""}, {"/url/", "a([a-z0-9]*)s([A-Z]{2})", "/to/{path}", ""},
...@@ -31,14 +31,17 @@ func TestRewrite(t *testing.T) { ...@@ -31,14 +31,17 @@ func TestRewrite(t *testing.T) {
{"/abcd/", "ab", "/a/{dir}/{file}", ".html|"}, {"/abcd/", "ab", "/a/{dir}/{file}", ".html|"},
{"/abcde/", "ab", "/a#{fragment}", ".html|"}, {"/abcde/", "ab", "/a#{fragment}", ".html|"},
{"/ab/", `.*\.jpg`, "/ajpg", ""}, {"/ab/", `.*\.jpg`, "/ajpg", ""},
{"/reggrp", `/ad/([0-9]+)([a-z]*)`, "/a{1}/{2}", ""},
{"/reg2grp", `(.*)`, "/{1}", ""},
{"/reg3grp", `(.*)/(.*)/(.*)`, "/{1}{2}{3}", ""},
} }
for _, regexpRule := range regexpRules { for _, regexpRule := range regexps {
var ext []string var ext []string
if s := strings.Split(regexpRule[3], "|"); len(s) > 1 { if s := strings.Split(regexpRule[3], "|"); len(s) > 1 {
ext = s[:len(s)-1] ext = s[:len(s)-1]
} }
rule, err := NewRegexpRule(regexpRule[0], regexpRule[1], regexpRule[2], ext) rule, err := NewComplexRule(regexpRule[0], regexpRule[1], regexpRule[2], ext, nil)
if err != nil { if err != nil {
t.Fatal(err) t.Fatal(err)
} }
...@@ -81,6 +84,12 @@ func TestRewrite(t *testing.T) { ...@@ -81,6 +84,12 @@ func TestRewrite(t *testing.T) {
{"/abcde/abcde.html", "/a"}, {"/abcde/abcde.html", "/a"},
{"/abcde/abcde.html#1234", "/a#1234"}, {"/abcde/abcde.html#1234", "/a#1234"},
{"/ab/ab.jpg", "/ajpg"}, {"/ab/ab.jpg", "/ajpg"},
{"/reggrp/ad/12", "/a12"},
{"/reggrp/ad/124a", "/a124/a"},
{"/reggrp/ad/124abc", "/a124/abc"},
{"/reg2grp/ad/124abc", "/ad/124abc"},
{"/reg3grp/ad/aa/66", "/adaa66"},
{"/reg3grp/ad612/n1n/ab", "/ad612n1nab"},
} }
for i, test := range tests { for i, test := range tests {
......
empty
\ No newline at end of file
package rewrite
import (
"log"
"net/http"
"net/url"
"path"
"strings"
"github.com/mholt/caddy/middleware"
)
// To attempts rewrite. It attempts to rewrite to first valid path
// or the last path if none of the paths are valid.
// Returns true if rewrite is successful and false otherwise.
func To(fs http.FileSystem, r *http.Request, to string, replacer middleware.Replacer) bool {
tos := strings.Fields(to)
// try each rewrite paths
t := ""
for _, v := range tos {
t = path.Clean(replacer.Replace(v))
// add trailing slash for directories, if present
if strings.HasSuffix(v, "/") && !strings.HasSuffix(t, "/") {
t += "/"
}
// validate file
if isValidFile(fs, t) {
break
}
}
// validate resulting path
u, err := url.Parse(t)
if err != nil {
// Let the user know we got here. Rewrite is expected but
// the resulting url is invalid.
log.Printf("[ERROR] rewrite: resulting path '%v' is invalid. error: %v", t, err)
return false
}
// take note of this rewrite for internal use by fastcgi
// all we need is the URI, not full URL
r.Header.Set(headerFieldName, r.URL.RequestURI())
// perform rewrite
r.URL.Path = u.Path
if u.RawQuery != "" {
// overwrite query string if present
r.URL.RawQuery = u.RawQuery
}
if u.Fragment != "" {
// overwrite fragment if present
r.URL.Fragment = u.Fragment
}
return true
}
// isValidFile checks if file exists on the filesystem.
// if file ends with `/`, it is validated as a directory.
func isValidFile(fs http.FileSystem, file string) bool {
if fs == nil {
return false
}
f, err := fs.Open(file)
if err != nil {
return false
}
defer f.Close()
stat, err := f.Stat()
if err != nil {
return false
}
// directory
if strings.HasSuffix(file, "/") {
return stat.IsDir()
}
// file
return !stat.IsDir()
}
package rewrite
import (
"net/http"
"net/url"
"testing"
)
func TestTo(t *testing.T) {
fs := http.Dir("testdata")
tests := []struct {
url string
to string
expected string
}{
{"/", "/somefiles", "/somefiles"},
{"/somefiles", "/somefiles /index.php{uri}", "/index.php/somefiles"},
{"/somefiles", "/testfile /index.php{uri}", "/testfile"},
{"/somefiles", "/testfile/ /index.php{uri}", "/index.php/somefiles"},
{"/somefiles", "/somefiles /index.php{uri}", "/index.php/somefiles"},
{"/?a=b", "/somefiles /index.php?{query}", "/index.php?a=b"},
{"/?a=b", "/testfile /index.php?{query}", "/testfile?a=b"},
{"/?a=b", "/testdir /index.php?{query}", "/index.php?a=b"},
{"/?a=b", "/testdir/ /index.php?{query}", "/testdir/?a=b"},
}
uri := func(r *url.URL) string {
uri := r.Path
if r.RawQuery != "" {
uri += "?" + r.RawQuery
}
return uri
}
for i, test := range tests {
r, err := http.NewRequest("GET", test.url, nil)
if err != nil {
t.Error(err)
}
To(fs, r, test.to, newReplacer(r))
if uri(r.URL) != test.expected {
t.Errorf("Test %v: expected %v found %v", i, test.expected, uri(r.URL))
}
}
}
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment