policies.go 16 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708
  1. // Copyright (c) 2012 The gocql Authors. All rights reserved.
  2. // Use of this source code is governed by a BSD-style
  3. // license that can be found in the LICENSE file.
  4. //This file will be the future home for more policies
  5. package gocql
  6. import (
  7. "fmt"
  8. "math"
  9. "math/rand"
  10. "net"
  11. "sync"
  12. "sync/atomic"
  13. "time"
  14. "github.com/hailocab/go-hostpool"
  15. )
  16. // cowHostList implements a copy on write host list, its equivalent type is []*HostInfo
  17. type cowHostList struct {
  18. list atomic.Value
  19. mu sync.Mutex
  20. }
  21. func (c *cowHostList) String() string {
  22. return fmt.Sprintf("%+v", c.get())
  23. }
  24. func (c *cowHostList) get() []*HostInfo {
  25. // TODO(zariel): should we replace this with []*HostInfo?
  26. l, ok := c.list.Load().(*[]*HostInfo)
  27. if !ok {
  28. return nil
  29. }
  30. return *l
  31. }
  32. func (c *cowHostList) set(list []*HostInfo) {
  33. c.mu.Lock()
  34. c.list.Store(&list)
  35. c.mu.Unlock()
  36. }
  37. // add will add a host if it not already in the list
  38. func (c *cowHostList) add(host *HostInfo) bool {
  39. c.mu.Lock()
  40. l := c.get()
  41. if n := len(l); n == 0 {
  42. l = []*HostInfo{host}
  43. } else {
  44. newL := make([]*HostInfo, n+1)
  45. for i := 0; i < n; i++ {
  46. if host.Equal(l[i]) {
  47. c.mu.Unlock()
  48. return false
  49. }
  50. newL[i] = l[i]
  51. }
  52. newL[n] = host
  53. l = newL
  54. }
  55. c.list.Store(&l)
  56. c.mu.Unlock()
  57. return true
  58. }
  59. func (c *cowHostList) update(host *HostInfo) {
  60. c.mu.Lock()
  61. l := c.get()
  62. if len(l) == 0 {
  63. c.mu.Unlock()
  64. return
  65. }
  66. found := false
  67. newL := make([]*HostInfo, len(l))
  68. for i := range l {
  69. if host.Equal(l[i]) {
  70. newL[i] = host
  71. found = true
  72. } else {
  73. newL[i] = l[i]
  74. }
  75. }
  76. if found {
  77. c.list.Store(&newL)
  78. }
  79. c.mu.Unlock()
  80. }
  81. func (c *cowHostList) remove(ip net.IP) bool {
  82. c.mu.Lock()
  83. l := c.get()
  84. size := len(l)
  85. if size == 0 {
  86. c.mu.Unlock()
  87. return false
  88. }
  89. found := false
  90. newL := make([]*HostInfo, 0, size)
  91. for i := 0; i < len(l); i++ {
  92. if !l[i].ConnectAddress().Equal(ip) {
  93. newL = append(newL, l[i])
  94. } else {
  95. found = true
  96. }
  97. }
  98. if !found {
  99. c.mu.Unlock()
  100. return false
  101. }
  102. newL = newL[:size-1 : size-1]
  103. c.list.Store(&newL)
  104. c.mu.Unlock()
  105. return true
  106. }
  107. // RetryableQuery is an interface that represents a query or batch statement that
  108. // exposes the correct functions for the retry policy logic to evaluate correctly.
  109. type RetryableQuery interface {
  110. Attempts() int
  111. GetConsistency() Consistency
  112. }
  113. // RetryPolicy interface is used by gocql to determine if a query can be attempted
  114. // again after a retryable error has been received. The interface allows gocql
  115. // users to implement their own logic to determine if a query can be attempted
  116. // again.
  117. //
  118. // See SimpleRetryPolicy as an example of implementing and using a RetryPolicy
  119. // interface.
  120. type RetryPolicy interface {
  121. Attempt(RetryableQuery) bool
  122. }
  123. // SimpleRetryPolicy has simple logic for attempting a query a fixed number of times.
  124. //
  125. // See below for examples of usage:
  126. //
  127. // //Assign to the cluster
  128. // cluster.RetryPolicy = &gocql.SimpleRetryPolicy{NumRetries: 3}
  129. //
  130. // //Assign to a query
  131. // query.RetryPolicy(&gocql.SimpleRetryPolicy{NumRetries: 1})
  132. //
  133. type SimpleRetryPolicy struct {
  134. NumRetries int //Number of times to retry a query
  135. }
  136. // Attempt tells gocql to attempt the query again based on query.Attempts being less
  137. // than the NumRetries defined in the policy.
  138. func (s *SimpleRetryPolicy) Attempt(q RetryableQuery) bool {
  139. return q.Attempts() <= s.NumRetries
  140. }
  141. // ExponentialBackoffRetryPolicy sleeps between attempts
  142. type ExponentialBackoffRetryPolicy struct {
  143. NumRetries int
  144. Min, Max time.Duration
  145. }
  146. func (e *ExponentialBackoffRetryPolicy) Attempt(q RetryableQuery) bool {
  147. if q.Attempts() > e.NumRetries {
  148. return false
  149. }
  150. time.Sleep(e.napTime(q.Attempts()))
  151. return true
  152. }
  153. func (e *ExponentialBackoffRetryPolicy) napTime(attempts int) time.Duration {
  154. if e.Min <= 0 {
  155. e.Min = 100 * time.Millisecond
  156. }
  157. if e.Max <= 0 {
  158. e.Max = 10 * time.Second
  159. }
  160. minFloat := float64(e.Min)
  161. napDuration := minFloat * math.Pow(2, float64(attempts-1))
  162. // add some jitter
  163. napDuration += rand.Float64()*minFloat - (minFloat / 2)
  164. if napDuration > float64(e.Max) {
  165. return time.Duration(e.Max)
  166. }
  167. return time.Duration(napDuration)
  168. }
  169. type HostStateNotifier interface {
  170. AddHost(host *HostInfo)
  171. RemoveHost(host *HostInfo)
  172. HostUp(host *HostInfo)
  173. HostDown(host *HostInfo)
  174. }
  175. type KeyspaceUpdateEvent struct {
  176. Keyspace string
  177. Change string
  178. }
  179. // HostSelectionPolicy is an interface for selecting
  180. // the most appropriate host to execute a given query.
  181. type HostSelectionPolicy interface {
  182. HostStateNotifier
  183. SetPartitioner
  184. KeyspaceChanged(KeyspaceUpdateEvent)
  185. Init(*Session)
  186. IsLocal(host *HostInfo) bool
  187. //Pick returns an iteration function over selected hosts
  188. Pick(ExecutableQuery) NextHost
  189. }
  190. // SelectedHost is an interface returned when picking a host from a host
  191. // selection policy.
  192. type SelectedHost interface {
  193. Info() *HostInfo
  194. Mark(error)
  195. }
  196. type selectedHost HostInfo
  197. func (host *selectedHost) Info() *HostInfo {
  198. return (*HostInfo)(host)
  199. }
  200. func (host *selectedHost) Mark(err error) {}
  201. // NextHost is an iteration function over picked hosts
  202. type NextHost func() SelectedHost
  203. // RoundRobinHostPolicy is a round-robin load balancing policy, where each host
  204. // is tried sequentially for each query.
  205. func RoundRobinHostPolicy() HostSelectionPolicy {
  206. return &roundRobinHostPolicy{}
  207. }
  208. type roundRobinHostPolicy struct {
  209. hosts cowHostList
  210. pos uint32
  211. mu sync.RWMutex
  212. }
  213. func (r *roundRobinHostPolicy) IsLocal(*HostInfo) bool { return true }
  214. func (r *roundRobinHostPolicy) KeyspaceChanged(KeyspaceUpdateEvent) {}
  215. func (r *roundRobinHostPolicy) SetPartitioner(partitioner string) {}
  216. func (r *roundRobinHostPolicy) Init(*Session) {}
  217. func (r *roundRobinHostPolicy) Pick(qry ExecutableQuery) NextHost {
  218. // i is used to limit the number of attempts to find a host
  219. // to the number of hosts known to this policy
  220. var i int
  221. return func() SelectedHost {
  222. hosts := r.hosts.get()
  223. if len(hosts) == 0 {
  224. return nil
  225. }
  226. // always increment pos to evenly distribute traffic in case of
  227. // failures
  228. pos := atomic.AddUint32(&r.pos, 1) - 1
  229. if i >= len(hosts) {
  230. return nil
  231. }
  232. host := hosts[(pos)%uint32(len(hosts))]
  233. i++
  234. return (*selectedHost)(host)
  235. }
  236. }
  237. func (r *roundRobinHostPolicy) AddHost(host *HostInfo) {
  238. r.hosts.add(host)
  239. }
  240. func (r *roundRobinHostPolicy) RemoveHost(host *HostInfo) {
  241. r.hosts.remove(host.ConnectAddress())
  242. }
  243. func (r *roundRobinHostPolicy) HostUp(host *HostInfo) {
  244. r.AddHost(host)
  245. }
  246. func (r *roundRobinHostPolicy) HostDown(host *HostInfo) {
  247. r.RemoveHost(host)
  248. }
  249. func ShuffleReplicas() func(*tokenAwareHostPolicy) {
  250. return func(t *tokenAwareHostPolicy) {
  251. t.shuffleReplicas = true
  252. }
  253. }
  254. // TokenAwareHostPolicy is a token aware host selection policy, where hosts are
  255. // selected based on the partition key, so queries are sent to the host which
  256. // owns the partition. Fallback is used when routing information is not available.
  257. func TokenAwareHostPolicy(fallback HostSelectionPolicy, opts ...func(*tokenAwareHostPolicy)) HostSelectionPolicy {
  258. p := &tokenAwareHostPolicy{fallback: fallback}
  259. for _, opt := range opts {
  260. opt(p)
  261. }
  262. return p
  263. }
  264. type keyspaceMeta struct {
  265. replicas map[string]map[token][]*HostInfo
  266. }
  267. type tokenAwareHostPolicy struct {
  268. hosts cowHostList
  269. mu sync.RWMutex
  270. partitioner string
  271. fallback HostSelectionPolicy
  272. session *Session
  273. tokenRing atomic.Value // *tokenRing
  274. keyspaces atomic.Value // *keyspaceMeta
  275. shuffleReplicas bool
  276. }
  277. func (t *tokenAwareHostPolicy) Init(s *Session) {
  278. t.session = s
  279. }
  280. func (t *tokenAwareHostPolicy) IsLocal(host *HostInfo) bool {
  281. return t.fallback.IsLocal(host)
  282. }
  283. func (t *tokenAwareHostPolicy) KeyspaceChanged(update KeyspaceUpdateEvent) {
  284. meta, _ := t.keyspaces.Load().(*keyspaceMeta)
  285. var size = 1
  286. if meta != nil {
  287. size = len(meta.replicas)
  288. }
  289. newMeta := &keyspaceMeta{
  290. replicas: make(map[string]map[token][]*HostInfo, size),
  291. }
  292. ks, err := t.session.KeyspaceMetadata(update.Keyspace)
  293. if err == nil {
  294. strat := getStrategy(ks)
  295. tr := t.tokenRing.Load().(*tokenRing)
  296. if tr != nil {
  297. newMeta.replicas[update.Keyspace] = strat.replicaMap(t.hosts.get(), tr.tokens)
  298. }
  299. }
  300. if meta != nil {
  301. for ks, replicas := range meta.replicas {
  302. if ks != update.Keyspace {
  303. newMeta.replicas[ks] = replicas
  304. }
  305. }
  306. }
  307. t.keyspaces.Store(newMeta)
  308. }
  309. func (t *tokenAwareHostPolicy) SetPartitioner(partitioner string) {
  310. t.mu.Lock()
  311. defer t.mu.Unlock()
  312. if t.partitioner != partitioner {
  313. t.fallback.SetPartitioner(partitioner)
  314. t.partitioner = partitioner
  315. t.resetTokenRing(partitioner)
  316. }
  317. }
  318. func (t *tokenAwareHostPolicy) AddHost(host *HostInfo) {
  319. t.hosts.add(host)
  320. t.fallback.AddHost(host)
  321. t.mu.RLock()
  322. partitioner := t.partitioner
  323. t.mu.RUnlock()
  324. t.resetTokenRing(partitioner)
  325. }
  326. func (t *tokenAwareHostPolicy) RemoveHost(host *HostInfo) {
  327. t.hosts.remove(host.ConnectAddress())
  328. t.fallback.RemoveHost(host)
  329. t.mu.RLock()
  330. partitioner := t.partitioner
  331. t.mu.RUnlock()
  332. t.resetTokenRing(partitioner)
  333. }
  334. func (t *tokenAwareHostPolicy) HostUp(host *HostInfo) {
  335. // TODO: need to avoid doing all the work on AddHost on hostup/down
  336. // because it now expensive to calculate the replica map for each
  337. // token
  338. t.AddHost(host)
  339. }
  340. func (t *tokenAwareHostPolicy) HostDown(host *HostInfo) {
  341. t.RemoveHost(host)
  342. }
  343. func (t *tokenAwareHostPolicy) resetTokenRing(partitioner string) {
  344. if partitioner == "" {
  345. // partitioner not yet set
  346. return
  347. }
  348. // create a new token ring
  349. hosts := t.hosts.get()
  350. tokenRing, err := newTokenRing(partitioner, hosts)
  351. if err != nil {
  352. Logger.Printf("Unable to update the token ring due to error: %s", err)
  353. return
  354. }
  355. // replace the token ring
  356. t.tokenRing.Store(tokenRing)
  357. }
  358. func (t *tokenAwareHostPolicy) getReplicas(keyspace string, token token) ([]*HostInfo, bool) {
  359. meta, _ := t.keyspaces.Load().(*keyspaceMeta)
  360. if meta == nil {
  361. return nil, false
  362. }
  363. tokens, ok := meta.replicas[keyspace][token]
  364. return tokens, ok
  365. }
  366. func (t *tokenAwareHostPolicy) Pick(qry ExecutableQuery) NextHost {
  367. if qry == nil {
  368. return t.fallback.Pick(qry)
  369. }
  370. routingKey, err := qry.GetRoutingKey()
  371. if err != nil {
  372. return t.fallback.Pick(qry)
  373. } else if routingKey == nil {
  374. return t.fallback.Pick(qry)
  375. }
  376. tr, _ := t.tokenRing.Load().(*tokenRing)
  377. if tr == nil {
  378. return t.fallback.Pick(qry)
  379. }
  380. token := tr.partitioner.Hash(routingKey)
  381. primaryEndpoint := tr.GetHostForToken(token)
  382. if primaryEndpoint == nil || token == nil {
  383. return t.fallback.Pick(qry)
  384. }
  385. replicas, ok := t.getReplicas(qry.Keyspace(), token)
  386. if !ok {
  387. replicas = []*HostInfo{primaryEndpoint}
  388. } else if t.shuffleReplicas {
  389. replicas = shuffleHosts(replicas)
  390. }
  391. var (
  392. fallbackIter NextHost
  393. i int
  394. )
  395. used := make(map[*HostInfo]bool, len(replicas))
  396. return func() SelectedHost {
  397. for i < len(replicas) {
  398. h := replicas[i]
  399. i++
  400. if h.IsUp() && t.fallback.IsLocal(h) {
  401. used[h] = true
  402. return (*selectedHost)(h)
  403. }
  404. }
  405. if fallbackIter == nil {
  406. // fallback
  407. fallbackIter = t.fallback.Pick(qry)
  408. }
  409. // filter the token aware selected hosts from the fallback hosts
  410. for fallbackHost := fallbackIter(); fallbackHost != nil; fallbackHost = fallbackIter() {
  411. if !used[fallbackHost.Info()] {
  412. return fallbackHost
  413. }
  414. }
  415. return nil
  416. }
  417. }
  418. // HostPoolHostPolicy is a host policy which uses the bitly/go-hostpool library
  419. // to distribute queries between hosts and prevent sending queries to
  420. // unresponsive hosts. When creating the host pool that is passed to the policy
  421. // use an empty slice of hosts as the hostpool will be populated later by gocql.
  422. // See below for examples of usage:
  423. //
  424. // // Create host selection policy using a simple host pool
  425. // cluster.PoolConfig.HostSelectionPolicy = HostPoolHostPolicy(hostpool.New(nil))
  426. //
  427. // // Create host selection policy using an epsilon greedy pool
  428. // cluster.PoolConfig.HostSelectionPolicy = HostPoolHostPolicy(
  429. // hostpool.NewEpsilonGreedy(nil, 0, &hostpool.LinearEpsilonValueCalculator{}),
  430. // )
  431. //
  432. func HostPoolHostPolicy(hp hostpool.HostPool) HostSelectionPolicy {
  433. return &hostPoolHostPolicy{hostMap: map[string]*HostInfo{}, hp: hp}
  434. }
  435. type hostPoolHostPolicy struct {
  436. hp hostpool.HostPool
  437. mu sync.RWMutex
  438. hostMap map[string]*HostInfo
  439. }
  440. func (r *hostPoolHostPolicy) Init(*Session) {}
  441. func (r *hostPoolHostPolicy) KeyspaceChanged(KeyspaceUpdateEvent) {}
  442. func (r *hostPoolHostPolicy) SetPartitioner(string) {}
  443. func (r *hostPoolHostPolicy) IsLocal(*HostInfo) bool { return true }
  444. func (r *hostPoolHostPolicy) SetHosts(hosts []*HostInfo) {
  445. peers := make([]string, len(hosts))
  446. hostMap := make(map[string]*HostInfo, len(hosts))
  447. for i, host := range hosts {
  448. ip := host.ConnectAddress().String()
  449. peers[i] = ip
  450. hostMap[ip] = host
  451. }
  452. r.mu.Lock()
  453. r.hp.SetHosts(peers)
  454. r.hostMap = hostMap
  455. r.mu.Unlock()
  456. }
  457. func (r *hostPoolHostPolicy) AddHost(host *HostInfo) {
  458. ip := host.ConnectAddress().String()
  459. r.mu.Lock()
  460. defer r.mu.Unlock()
  461. // If the host addr is present and isn't nil return
  462. if h, ok := r.hostMap[ip]; ok && h != nil {
  463. return
  464. }
  465. // otherwise, add the host to the map
  466. r.hostMap[ip] = host
  467. // and construct a new peer list to give to the HostPool
  468. hosts := make([]string, 0, len(r.hostMap))
  469. for addr := range r.hostMap {
  470. hosts = append(hosts, addr)
  471. }
  472. r.hp.SetHosts(hosts)
  473. }
  474. func (r *hostPoolHostPolicy) RemoveHost(host *HostInfo) {
  475. ip := host.ConnectAddress().String()
  476. r.mu.Lock()
  477. defer r.mu.Unlock()
  478. if _, ok := r.hostMap[ip]; !ok {
  479. return
  480. }
  481. delete(r.hostMap, ip)
  482. hosts := make([]string, 0, len(r.hostMap))
  483. for _, host := range r.hostMap {
  484. hosts = append(hosts, host.ConnectAddress().String())
  485. }
  486. r.hp.SetHosts(hosts)
  487. }
  488. func (r *hostPoolHostPolicy) HostUp(host *HostInfo) {
  489. r.AddHost(host)
  490. }
  491. func (r *hostPoolHostPolicy) HostDown(host *HostInfo) {
  492. r.RemoveHost(host)
  493. }
  494. func (r *hostPoolHostPolicy) Pick(qry ExecutableQuery) NextHost {
  495. return func() SelectedHost {
  496. r.mu.RLock()
  497. defer r.mu.RUnlock()
  498. if len(r.hostMap) == 0 {
  499. return nil
  500. }
  501. hostR := r.hp.Get()
  502. host, ok := r.hostMap[hostR.Host()]
  503. if !ok {
  504. return nil
  505. }
  506. return selectedHostPoolHost{
  507. policy: r,
  508. info: host,
  509. hostR: hostR,
  510. }
  511. }
  512. }
  513. // selectedHostPoolHost is a host returned by the hostPoolHostPolicy and
  514. // implements the SelectedHost interface
  515. type selectedHostPoolHost struct {
  516. policy *hostPoolHostPolicy
  517. info *HostInfo
  518. hostR hostpool.HostPoolResponse
  519. }
  520. func (host selectedHostPoolHost) Info() *HostInfo {
  521. return host.info
  522. }
  523. func (host selectedHostPoolHost) Mark(err error) {
  524. ip := host.info.ConnectAddress().String()
  525. host.policy.mu.RLock()
  526. defer host.policy.mu.RUnlock()
  527. if _, ok := host.policy.hostMap[ip]; !ok {
  528. // host was removed between pick and mark
  529. return
  530. }
  531. host.hostR.Mark(err)
  532. }
  533. type dcAwareRR struct {
  534. local string
  535. pos uint32
  536. mu sync.RWMutex
  537. localHosts cowHostList
  538. remoteHosts cowHostList
  539. }
  540. // DCAwareRoundRobinPolicy is a host selection policies which will prioritize and
  541. // return hosts which are in the local datacentre before returning hosts in all
  542. // other datercentres
  543. func DCAwareRoundRobinPolicy(localDC string) HostSelectionPolicy {
  544. return &dcAwareRR{local: localDC}
  545. }
  546. func (d *dcAwareRR) Init(*Session) {}
  547. func (d *dcAwareRR) KeyspaceChanged(KeyspaceUpdateEvent) {}
  548. func (d *dcAwareRR) SetPartitioner(p string) {}
  549. func (d *dcAwareRR) IsLocal(host *HostInfo) bool {
  550. return host.DataCenter() == d.local
  551. }
  552. func (d *dcAwareRR) AddHost(host *HostInfo) {
  553. if host.DataCenter() == d.local {
  554. d.localHosts.add(host)
  555. } else {
  556. d.remoteHosts.add(host)
  557. }
  558. }
  559. func (d *dcAwareRR) RemoveHost(host *HostInfo) {
  560. if host.DataCenter() == d.local {
  561. d.localHosts.remove(host.ConnectAddress())
  562. } else {
  563. d.remoteHosts.remove(host.ConnectAddress())
  564. }
  565. }
  566. func (d *dcAwareRR) HostUp(host *HostInfo) { d.AddHost(host) }
  567. func (d *dcAwareRR) HostDown(host *HostInfo) { d.RemoveHost(host) }
  568. func (d *dcAwareRR) Pick(q ExecutableQuery) NextHost {
  569. var i int
  570. return func() SelectedHost {
  571. var hosts []*HostInfo
  572. localHosts := d.localHosts.get()
  573. remoteHosts := d.remoteHosts.get()
  574. if len(localHosts) != 0 {
  575. hosts = localHosts
  576. } else {
  577. hosts = remoteHosts
  578. }
  579. if len(hosts) == 0 {
  580. return nil
  581. }
  582. // always increment pos to evenly distribute traffic in case of
  583. // failures
  584. pos := atomic.AddUint32(&d.pos, 1) - 1
  585. if i >= len(localHosts)+len(remoteHosts) {
  586. return nil
  587. }
  588. host := hosts[(pos)%uint32(len(hosts))]
  589. i++
  590. return (*selectedHost)(host)
  591. }
  592. }