| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366 |
- package httprule
- import (
- "flag"
- "fmt"
- "reflect"
- "testing"
- "github.com/golang/glog"
- )
- func TestTokenize(t *testing.T) {
- for _, spec := range []struct {
- src string
- tokens []string
- verb string
- }{
- {
- src: "",
- tokens: []string{eof},
- },
- {
- src: "v1",
- tokens: []string{"v1", eof},
- },
- {
- src: "v1/b",
- tokens: []string{"v1", "/", "b", eof},
- },
- {
- src: "v1/endpoint/*",
- tokens: []string{"v1", "/", "endpoint", "/", "*", eof},
- },
- {
- src: "v1/endpoint/**",
- tokens: []string{"v1", "/", "endpoint", "/", "**", eof},
- },
- {
- src: "v1/b/{bucket_name=*}",
- tokens: []string{
- "v1", "/",
- "b", "/",
- "{", "bucket_name", "=", "*", "}",
- eof,
- },
- },
- {
- src: "v1/b/{bucket_name=buckets/*}",
- tokens: []string{
- "v1", "/",
- "b", "/",
- "{", "bucket_name", "=", "buckets", "/", "*", "}",
- eof,
- },
- },
- {
- src: "v1/b/{bucket_name=buckets/*}/o",
- tokens: []string{
- "v1", "/",
- "b", "/",
- "{", "bucket_name", "=", "buckets", "/", "*", "}", "/",
- "o",
- eof,
- },
- },
- {
- src: "v1/b/{bucket_name=buckets/*}/o/{name}",
- tokens: []string{
- "v1", "/",
- "b", "/",
- "{", "bucket_name", "=", "buckets", "/", "*", "}", "/",
- "o", "/", "{", "name", "}",
- eof,
- },
- },
- {
- src: "v1/a=b&c=d;e=f:g/endpoint.rdf",
- tokens: []string{
- "v1", "/",
- "a=b&c=d;e=f:g", "/",
- "endpoint.rdf",
- eof,
- },
- },
- {
- src: "v1/a/{endpoint}:a",
- tokens: []string{
- "v1", "/",
- "a", "/",
- "{", "endpoint", "}",
- eof,
- },
- verb: "a",
- },
- {
- src: "v1/a/{endpoint}:b:c",
- tokens: []string{
- "v1", "/",
- "a", "/",
- "{", "endpoint", "}",
- eof,
- },
- verb: "b:c",
- },
- } {
- tokens, verb := tokenize(spec.src)
- if got, want := tokens, spec.tokens; !reflect.DeepEqual(got, want) {
- t.Errorf("tokenize(%q) = %q, _; want %q, _", spec.src, got, want)
- }
- switch {
- case spec.verb != "":
- if got, want := verb, spec.verb; !reflect.DeepEqual(got, want) {
- t.Errorf("tokenize(%q) = %q, _; want %q, _", spec.src, got, want)
- }
- default:
- if got, want := verb, ""; got != want {
- t.Errorf("tokenize(%q) = _, %q; want _, %q", spec.src, got, want)
- }
- src := fmt.Sprintf("%s:%s", spec.src, "LOCK")
- tokens, verb = tokenize(src)
- if got, want := tokens, spec.tokens; !reflect.DeepEqual(got, want) {
- t.Errorf("tokenize(%q) = %q, _; want %q, _", src, got, want)
- }
- if got, want := verb, "LOCK"; got != want {
- t.Errorf("tokenize(%q) = _, %q; want _, %q", src, got, want)
- }
- }
- }
- }
- func TestParseSegments(t *testing.T) {
- err := flag.Set("v", "3")
- if err != nil {
- t.Fatalf("failed to set flag: %v", err)
- }
- for _, spec := range []struct {
- tokens []string
- want []segment
- }{
- {
- tokens: []string{eof},
- want: []segment{
- literal(eof),
- },
- },
- {
- // Note: this case will never arise as tokenize() will never return such a sequence of tokens
- // and even if it does it will be treated as [eof]
- tokens: []string{eof, "v1", eof},
- want: []segment{
- literal(eof),
- },
- },
- {
- tokens: []string{"v1", eof},
- want: []segment{
- literal("v1"),
- },
- },
- {
- tokens: []string{"/", eof},
- want: []segment{
- wildcard{},
- },
- },
- {
- tokens: []string{"-._~!$&'()*+,;=:@", eof},
- want: []segment{
- literal("-._~!$&'()*+,;=:@"),
- },
- },
- {
- tokens: []string{"%e7%ac%ac%e4%b8%80%e7%89%88", eof},
- want: []segment{
- literal("%e7%ac%ac%e4%b8%80%e7%89%88"),
- },
- },
- {
- tokens: []string{"v1", "/", "*", eof},
- want: []segment{
- literal("v1"),
- wildcard{},
- },
- },
- {
- tokens: []string{"v1", "/", "**", eof},
- want: []segment{
- literal("v1"),
- deepWildcard{},
- },
- },
- {
- tokens: []string{"{", "name", "}", eof},
- want: []segment{
- variable{
- path: "name",
- segments: []segment{
- wildcard{},
- },
- },
- },
- },
- {
- tokens: []string{"{", "name", "=", "*", "}", eof},
- want: []segment{
- variable{
- path: "name",
- segments: []segment{
- wildcard{},
- },
- },
- },
- },
- {
- tokens: []string{"{", "field", ".", "nested", ".", "nested2", "=", "*", "}", eof},
- want: []segment{
- variable{
- path: "field.nested.nested2",
- segments: []segment{
- wildcard{},
- },
- },
- },
- },
- {
- tokens: []string{"{", "name", "=", "a", "/", "b", "/", "*", "}", eof},
- want: []segment{
- variable{
- path: "name",
- segments: []segment{
- literal("a"),
- literal("b"),
- wildcard{},
- },
- },
- },
- },
- {
- tokens: []string{
- "v1", "/",
- "{",
- "name", ".", "nested", ".", "nested2",
- "=",
- "a", "/", "b", "/", "*",
- "}", "/",
- "o", "/",
- "{",
- "another_name",
- "=",
- "a", "/", "b", "/", "*", "/", "c",
- "}", "/",
- "**",
- eof,
- },
- want: []segment{
- literal("v1"),
- variable{
- path: "name.nested.nested2",
- segments: []segment{
- literal("a"),
- literal("b"),
- wildcard{},
- },
- },
- literal("o"),
- variable{
- path: "another_name",
- segments: []segment{
- literal("a"),
- literal("b"),
- wildcard{},
- literal("c"),
- },
- },
- deepWildcard{},
- },
- },
- } {
- p := parser{tokens: spec.tokens}
- segs, err := p.topLevelSegments()
- if err != nil {
- t.Errorf("parser{%q}.segments() failed with %v; want success", spec.tokens, err)
- continue
- }
- if got, want := segs, spec.want; !reflect.DeepEqual(got, want) {
- t.Errorf("parser{%q}.segments() = %#v; want %#v", spec.tokens, got, want)
- }
- if got := p.tokens; len(got) > 0 {
- t.Errorf("p.tokens = %q; want []; spec.tokens=%q", got, spec.tokens)
- }
- }
- }
- func TestParseSegmentsWithErrors(t *testing.T) {
- err := flag.Set("v", "3")
- if err != nil {
- t.Fatalf("failed to set flag: %v", err)
- }
- for _, spec := range []struct {
- tokens []string
- }{
- {
- // double slash
- tokens: []string{"//", eof},
- },
- {
- // invalid literal
- tokens: []string{"a?b", eof},
- },
- {
- // invalid percent-encoding
- tokens: []string{"%", eof},
- },
- {
- // invalid percent-encoding
- tokens: []string{"%2", eof},
- },
- {
- // invalid percent-encoding
- tokens: []string{"a%2z", eof},
- },
- {
- // unterminated variable
- tokens: []string{"{", "name", eof},
- },
- {
- // unterminated variable
- tokens: []string{"{", "name", "=", eof},
- },
- {
- // unterminated variable
- tokens: []string{"{", "name", "=", "*", eof},
- },
- {
- // empty component in field path
- tokens: []string{"{", "name", ".", "}", eof},
- },
- {
- // empty component in field path
- tokens: []string{"{", "name", ".", ".", "nested", "}", eof},
- },
- {
- // invalid character in identifier
- tokens: []string{"{", "field-name", "}", eof},
- },
- {
- // no slash between segments
- tokens: []string{"v1", "endpoint", eof},
- },
- {
- // no slash between segments
- tokens: []string{"v1", "{", "name", "}", eof},
- },
- } {
- p := parser{tokens: spec.tokens}
- segs, err := p.topLevelSegments()
- if err == nil {
- t.Errorf("parser{%q}.segments() succeeded; want InvalidTemplateError; accepted %#v", spec.tokens, segs)
- continue
- }
- glog.V(1).Info(err)
- }
- }
|