parse_test.go 7.0 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366
  1. package httprule
  2. import (
  3. "flag"
  4. "fmt"
  5. "reflect"
  6. "testing"
  7. "github.com/golang/glog"
  8. )
  9. func TestTokenize(t *testing.T) {
  10. for _, spec := range []struct {
  11. src string
  12. tokens []string
  13. verb string
  14. }{
  15. {
  16. src: "",
  17. tokens: []string{eof},
  18. },
  19. {
  20. src: "v1",
  21. tokens: []string{"v1", eof},
  22. },
  23. {
  24. src: "v1/b",
  25. tokens: []string{"v1", "/", "b", eof},
  26. },
  27. {
  28. src: "v1/endpoint/*",
  29. tokens: []string{"v1", "/", "endpoint", "/", "*", eof},
  30. },
  31. {
  32. src: "v1/endpoint/**",
  33. tokens: []string{"v1", "/", "endpoint", "/", "**", eof},
  34. },
  35. {
  36. src: "v1/b/{bucket_name=*}",
  37. tokens: []string{
  38. "v1", "/",
  39. "b", "/",
  40. "{", "bucket_name", "=", "*", "}",
  41. eof,
  42. },
  43. },
  44. {
  45. src: "v1/b/{bucket_name=buckets/*}",
  46. tokens: []string{
  47. "v1", "/",
  48. "b", "/",
  49. "{", "bucket_name", "=", "buckets", "/", "*", "}",
  50. eof,
  51. },
  52. },
  53. {
  54. src: "v1/b/{bucket_name=buckets/*}/o",
  55. tokens: []string{
  56. "v1", "/",
  57. "b", "/",
  58. "{", "bucket_name", "=", "buckets", "/", "*", "}", "/",
  59. "o",
  60. eof,
  61. },
  62. },
  63. {
  64. src: "v1/b/{bucket_name=buckets/*}/o/{name}",
  65. tokens: []string{
  66. "v1", "/",
  67. "b", "/",
  68. "{", "bucket_name", "=", "buckets", "/", "*", "}", "/",
  69. "o", "/", "{", "name", "}",
  70. eof,
  71. },
  72. },
  73. {
  74. src: "v1/a=b&c=d;e=f:g/endpoint.rdf",
  75. tokens: []string{
  76. "v1", "/",
  77. "a=b&c=d;e=f:g", "/",
  78. "endpoint.rdf",
  79. eof,
  80. },
  81. },
  82. {
  83. src: "v1/a/{endpoint}:a",
  84. tokens: []string{
  85. "v1", "/",
  86. "a", "/",
  87. "{", "endpoint", "}",
  88. eof,
  89. },
  90. verb: "a",
  91. },
  92. {
  93. src: "v1/a/{endpoint}:b:c",
  94. tokens: []string{
  95. "v1", "/",
  96. "a", "/",
  97. "{", "endpoint", "}",
  98. eof,
  99. },
  100. verb: "b:c",
  101. },
  102. } {
  103. tokens, verb := tokenize(spec.src)
  104. if got, want := tokens, spec.tokens; !reflect.DeepEqual(got, want) {
  105. t.Errorf("tokenize(%q) = %q, _; want %q, _", spec.src, got, want)
  106. }
  107. switch {
  108. case spec.verb != "":
  109. if got, want := verb, spec.verb; !reflect.DeepEqual(got, want) {
  110. t.Errorf("tokenize(%q) = %q, _; want %q, _", spec.src, got, want)
  111. }
  112. default:
  113. if got, want := verb, ""; got != want {
  114. t.Errorf("tokenize(%q) = _, %q; want _, %q", spec.src, got, want)
  115. }
  116. src := fmt.Sprintf("%s:%s", spec.src, "LOCK")
  117. tokens, verb = tokenize(src)
  118. if got, want := tokens, spec.tokens; !reflect.DeepEqual(got, want) {
  119. t.Errorf("tokenize(%q) = %q, _; want %q, _", src, got, want)
  120. }
  121. if got, want := verb, "LOCK"; got != want {
  122. t.Errorf("tokenize(%q) = _, %q; want _, %q", src, got, want)
  123. }
  124. }
  125. }
  126. }
  127. func TestParseSegments(t *testing.T) {
  128. err := flag.Set("v", "3")
  129. if err != nil {
  130. t.Fatalf("failed to set flag: %v", err)
  131. }
  132. for _, spec := range []struct {
  133. tokens []string
  134. want []segment
  135. }{
  136. {
  137. tokens: []string{eof},
  138. want: []segment{
  139. literal(eof),
  140. },
  141. },
  142. {
  143. // Note: this case will never arise as tokenize() will never return such a sequence of tokens
  144. // and even if it does it will be treated as [eof]
  145. tokens: []string{eof, "v1", eof},
  146. want: []segment{
  147. literal(eof),
  148. },
  149. },
  150. {
  151. tokens: []string{"v1", eof},
  152. want: []segment{
  153. literal("v1"),
  154. },
  155. },
  156. {
  157. tokens: []string{"/", eof},
  158. want: []segment{
  159. wildcard{},
  160. },
  161. },
  162. {
  163. tokens: []string{"-._~!$&'()*+,;=:@", eof},
  164. want: []segment{
  165. literal("-._~!$&'()*+,;=:@"),
  166. },
  167. },
  168. {
  169. tokens: []string{"%e7%ac%ac%e4%b8%80%e7%89%88", eof},
  170. want: []segment{
  171. literal("%e7%ac%ac%e4%b8%80%e7%89%88"),
  172. },
  173. },
  174. {
  175. tokens: []string{"v1", "/", "*", eof},
  176. want: []segment{
  177. literal("v1"),
  178. wildcard{},
  179. },
  180. },
  181. {
  182. tokens: []string{"v1", "/", "**", eof},
  183. want: []segment{
  184. literal("v1"),
  185. deepWildcard{},
  186. },
  187. },
  188. {
  189. tokens: []string{"{", "name", "}", eof},
  190. want: []segment{
  191. variable{
  192. path: "name",
  193. segments: []segment{
  194. wildcard{},
  195. },
  196. },
  197. },
  198. },
  199. {
  200. tokens: []string{"{", "name", "=", "*", "}", eof},
  201. want: []segment{
  202. variable{
  203. path: "name",
  204. segments: []segment{
  205. wildcard{},
  206. },
  207. },
  208. },
  209. },
  210. {
  211. tokens: []string{"{", "field", ".", "nested", ".", "nested2", "=", "*", "}", eof},
  212. want: []segment{
  213. variable{
  214. path: "field.nested.nested2",
  215. segments: []segment{
  216. wildcard{},
  217. },
  218. },
  219. },
  220. },
  221. {
  222. tokens: []string{"{", "name", "=", "a", "/", "b", "/", "*", "}", eof},
  223. want: []segment{
  224. variable{
  225. path: "name",
  226. segments: []segment{
  227. literal("a"),
  228. literal("b"),
  229. wildcard{},
  230. },
  231. },
  232. },
  233. },
  234. {
  235. tokens: []string{
  236. "v1", "/",
  237. "{",
  238. "name", ".", "nested", ".", "nested2",
  239. "=",
  240. "a", "/", "b", "/", "*",
  241. "}", "/",
  242. "o", "/",
  243. "{",
  244. "another_name",
  245. "=",
  246. "a", "/", "b", "/", "*", "/", "c",
  247. "}", "/",
  248. "**",
  249. eof,
  250. },
  251. want: []segment{
  252. literal("v1"),
  253. variable{
  254. path: "name.nested.nested2",
  255. segments: []segment{
  256. literal("a"),
  257. literal("b"),
  258. wildcard{},
  259. },
  260. },
  261. literal("o"),
  262. variable{
  263. path: "another_name",
  264. segments: []segment{
  265. literal("a"),
  266. literal("b"),
  267. wildcard{},
  268. literal("c"),
  269. },
  270. },
  271. deepWildcard{},
  272. },
  273. },
  274. } {
  275. p := parser{tokens: spec.tokens}
  276. segs, err := p.topLevelSegments()
  277. if err != nil {
  278. t.Errorf("parser{%q}.segments() failed with %v; want success", spec.tokens, err)
  279. continue
  280. }
  281. if got, want := segs, spec.want; !reflect.DeepEqual(got, want) {
  282. t.Errorf("parser{%q}.segments() = %#v; want %#v", spec.tokens, got, want)
  283. }
  284. if got := p.tokens; len(got) > 0 {
  285. t.Errorf("p.tokens = %q; want []; spec.tokens=%q", got, spec.tokens)
  286. }
  287. }
  288. }
  289. func TestParseSegmentsWithErrors(t *testing.T) {
  290. err := flag.Set("v", "3")
  291. if err != nil {
  292. t.Fatalf("failed to set flag: %v", err)
  293. }
  294. for _, spec := range []struct {
  295. tokens []string
  296. }{
  297. {
  298. // double slash
  299. tokens: []string{"//", eof},
  300. },
  301. {
  302. // invalid literal
  303. tokens: []string{"a?b", eof},
  304. },
  305. {
  306. // invalid percent-encoding
  307. tokens: []string{"%", eof},
  308. },
  309. {
  310. // invalid percent-encoding
  311. tokens: []string{"%2", eof},
  312. },
  313. {
  314. // invalid percent-encoding
  315. tokens: []string{"a%2z", eof},
  316. },
  317. {
  318. // unterminated variable
  319. tokens: []string{"{", "name", eof},
  320. },
  321. {
  322. // unterminated variable
  323. tokens: []string{"{", "name", "=", eof},
  324. },
  325. {
  326. // unterminated variable
  327. tokens: []string{"{", "name", "=", "*", eof},
  328. },
  329. {
  330. // empty component in field path
  331. tokens: []string{"{", "name", ".", "}", eof},
  332. },
  333. {
  334. // empty component in field path
  335. tokens: []string{"{", "name", ".", ".", "nested", "}", eof},
  336. },
  337. {
  338. // invalid character in identifier
  339. tokens: []string{"{", "field-name", "}", eof},
  340. },
  341. {
  342. // no slash between segments
  343. tokens: []string{"v1", "endpoint", eof},
  344. },
  345. {
  346. // no slash between segments
  347. tokens: []string{"v1", "{", "name", "}", eof},
  348. },
  349. } {
  350. p := parser{tokens: spec.tokens}
  351. segs, err := p.topLevelSegments()
  352. if err == nil {
  353. t.Errorf("parser{%q}.segments() succeeded; want InvalidTemplateError; accepted %#v", spec.tokens, segs)
  354. continue
  355. }
  356. glog.V(1).Info(err)
  357. }
  358. }