Вы не можете выбрать более 25 тем Темы должны начинаться с буквы или цифры, могут содержать дефисы(-) и должны содержать не более 35 символов.

542 строки
14 KiB

  1. package main
  2. import (
  3. "bufio"
  4. "bytes"
  5. "context"
  6. "encoding/json"
  7. "fmt"
  8. "io"
  9. "log"
  10. "net/http"
  11. "os"
  12. "os/signal"
  13. "strings"
  14. "sync"
  15. "syscall"
  16. "time"
  17. "github.com/go-redis/redis/v8"
  18. "github.com/gorilla/mux"
  19. )
  20. const (
  21. ItemChannelBuffer = 100000
  22. ItemWrapSize = 100000
  23. )
  24. type ProjectRedisConfig struct {
  25. Host string `json:"host"`
  26. Pass string `json:"pass"`
  27. Port int `json:"port"`
  28. }
  29. type ProjectConfig struct {
  30. RedisConfig *ProjectRedisConfig `json:"redis,omitempty"`
  31. }
  32. type BackfeedItem struct {
  33. PrimaryShard byte
  34. SecondaryShard string
  35. Item []byte
  36. }
  37. type ProjectBackfeedManager struct {
  38. Context context.Context
  39. Cancel context.CancelFunc
  40. Done chan bool
  41. C chan *BackfeedItem
  42. Name string
  43. BackfeedRedis *redis.ClusterClient
  44. ProjectRedis *redis.Client
  45. LegacyRedis *redis.Client
  46. Lock sync.RWMutex
  47. ProjectConfig ProjectConfig
  48. }
  49. func (that *ProjectBackfeedManager) RedisConfigDiffers(new *ProjectRedisConfig) bool {
  50. if that.ProjectConfig.RedisConfig == nil && new == nil {
  51. return false
  52. }
  53. return that.ProjectConfig.RedisConfig == nil || new == nil || *that.ProjectConfig.RedisConfig != *new
  54. }
  55. func (that *ProjectBackfeedManager) PushItem(ctx context.Context, item *BackfeedItem) bool {
  56. that.Lock.RLock()
  57. defer that.Lock.RUnlock()
  58. if that.C == nil {
  59. return false
  60. }
  61. select {
  62. case <-ctx.Done():
  63. return false
  64. case <-that.Context.Done():
  65. return false
  66. case that.C <- item:
  67. return true
  68. }
  69. }
  70. func (that *ProjectBackfeedManager) PopItem(blocking bool) (*BackfeedItem, bool) {
  71. if blocking {
  72. select {
  73. case <-that.Context.Done():
  74. return nil, false
  75. case item := <-that.C:
  76. return item, true
  77. }
  78. } else {
  79. select {
  80. case <-that.Context.Done():
  81. return nil, false
  82. case item := <-that.C:
  83. return item, true
  84. default:
  85. return nil, false
  86. }
  87. }
  88. }
  89. func (that *ProjectBackfeedManager) CloseItemChannel() {
  90. that.Lock.Lock()
  91. defer that.Lock.Unlock()
  92. if that.C == nil {
  93. return
  94. }
  95. close(that.C)
  96. that.C = nil
  97. }
  98. func (that *ProjectBackfeedManager) Do() {
  99. defer close(that.Done)
  100. defer that.CloseItemChannel()
  101. defer that.Cancel()
  102. for {
  103. select {
  104. case <-that.Context.Done():
  105. break
  106. case <-that.Done:
  107. break
  108. default:
  109. }
  110. item, ok := that.PopItem(true)
  111. if !ok {
  112. break
  113. }
  114. keyMap := map[string][][]byte{}
  115. key := fmt.Sprintf("%s:%02x:%s", that.Name, item.PrimaryShard, item.SecondaryShard)
  116. keyMap[key] = append(keyMap[key], item.Item)
  117. wrapped := 1
  118. for wrapped < ItemWrapSize {
  119. item, ok := that.PopItem(false)
  120. if !ok {
  121. break
  122. }
  123. key := fmt.Sprintf("%s:%02x:%s", that.Name, item.PrimaryShard, item.SecondaryShard)
  124. keyMap[key] = append(keyMap[key], item.Item)
  125. wrapped++
  126. }
  127. select {
  128. case <-that.Context.Done():
  129. break
  130. case <-that.Done:
  131. break
  132. default:
  133. }
  134. resultMap := map[string]*redis.Cmd{}
  135. pipe := that.BackfeedRedis.Pipeline()
  136. for key, items := range keyMap {
  137. args := []interface{}{
  138. "bf.madd",
  139. key,
  140. }
  141. for _, item := range items {
  142. args = append(args, item)
  143. }
  144. resultMap[key] = pipe.Do(context.Background(), args...)
  145. }
  146. _, err := pipe.Exec(context.Background())
  147. if err != nil {
  148. log.Printf("%s", err)
  149. continue
  150. }
  151. var sAddItems []interface{}
  152. for key, items := range keyMap {
  153. res, err := resultMap[key].BoolSlice()
  154. if err != nil {
  155. log.Printf("%s", err)
  156. continue
  157. }
  158. if len(res) != len(keyMap[key]) {
  159. continue
  160. }
  161. for i, v := range res {
  162. if v {
  163. sAddItems = append(sAddItems, items[i])
  164. }
  165. }
  166. }
  167. dupes := wrapped - len(sAddItems)
  168. if len(sAddItems) != 0 {
  169. args := []interface{}{
  170. "bf.mexists",
  171. that.Name,
  172. }
  173. args = append(args, sAddItems...)
  174. res, err := that.LegacyRedis.Do(context.Background(), args...).BoolSlice()
  175. if err != nil {
  176. log.Printf("unable to dedupe against %s legacy backfeed: %s", that.Name, err)
  177. } else if len(res) == len(sAddItems) {
  178. var filteredSAddItems []interface{}
  179. for i, v := range res {
  180. if !v {
  181. filteredSAddItems = append(filteredSAddItems, sAddItems[i])
  182. }
  183. }
  184. sAddItems = filteredSAddItems
  185. }
  186. }
  187. if len(sAddItems) != 0 {
  188. err := that.ProjectRedis.SAdd(context.Background(), fmt.Sprintf("%s:todo:backfeed", that.Name), sAddItems...).Err()
  189. if err != nil {
  190. log.Printf("failed to sadd items for %s: %s", that.Name, err)
  191. }
  192. }
  193. if dupes > 0 {
  194. that.BackfeedRedis.HIncrBy(context.Background(), ":", that.Name, int64(dupes))
  195. }
  196. }
  197. }
  198. type GlobalBackfeedManager struct {
  199. Context context.Context
  200. Cancel context.CancelFunc
  201. ActiveFeeds map[string]*ProjectBackfeedManager
  202. ActiveSlugs map[string]string
  203. TrackerRedis *redis.Client
  204. BackfeedRedis *redis.ClusterClient
  205. LegacyRedis *redis.Client
  206. Lock sync.RWMutex
  207. }
  208. func (that *GlobalBackfeedManager) RefreshFeeds() error {
  209. slugProjectMap, err := that.TrackerRedis.HGetAll(that.Context, "backfeed").Result()
  210. if err != nil {
  211. return err
  212. }
  213. var projects []string
  214. projectSlugMap := map[string][]string{}
  215. for slug, project := range slugProjectMap {
  216. projectSlugMap[project] = append(projectSlugMap[project], slug)
  217. }
  218. for project := range projectSlugMap {
  219. projects = append(projects, project)
  220. }
  221. projectConfigs := map[string]ProjectConfig{}
  222. if len(projects) != 0 {
  223. cfgi, err := that.TrackerRedis.HMGet(that.Context, "trackers", projects...).Result()
  224. if err != nil {
  225. return err
  226. }
  227. if len(projects) != len(cfgi) {
  228. return fmt.Errorf("hmget result had unexpected length")
  229. }
  230. for i, project := range projects {
  231. configString, ok := cfgi[i].(string)
  232. if !ok {
  233. continue
  234. }
  235. config := ProjectConfig{}
  236. err := json.Unmarshal([]byte(configString), &config)
  237. if err != nil {
  238. continue
  239. }
  240. projectConfigs[project] = config
  241. }
  242. }
  243. projects = nil
  244. for project := range projectSlugMap {
  245. if _, has := projectConfigs[project]; !has {
  246. delete(projectSlugMap, project)
  247. continue
  248. }
  249. projects = append(projects, project)
  250. }
  251. for slug, project := range slugProjectMap {
  252. if _, has := projectConfigs[project]; !has {
  253. delete(slugProjectMap, slug)
  254. }
  255. }
  256. // add feeds for new projects
  257. for _, project := range projects {
  258. projectConfig := projectConfigs[project]
  259. var outdatedProjectBackfeedManager *ProjectBackfeedManager
  260. if projectBackfeedManager, has := that.ActiveFeeds[project]; has {
  261. if that.ActiveFeeds[project].RedisConfigDiffers(projectConfig.RedisConfig) {
  262. outdatedProjectBackfeedManager = projectBackfeedManager
  263. } else {
  264. continue
  265. }
  266. }
  267. ctx, cancel := context.WithCancel(that.Context)
  268. projectBackfeedManager := &ProjectBackfeedManager{
  269. Context: ctx,
  270. Cancel: cancel,
  271. Done: make(chan bool),
  272. C: make(chan *BackfeedItem, ItemChannelBuffer),
  273. BackfeedRedis: that.BackfeedRedis,
  274. Name: project,
  275. ProjectConfig: projectConfig,
  276. LegacyRedis: that.LegacyRedis,
  277. }
  278. if projectConfig.RedisConfig != nil {
  279. projectBackfeedManager.ProjectRedis = redis.NewClient(&redis.Options{
  280. Addr: fmt.Sprintf("%s:%d", projectConfig.RedisConfig.Host, projectConfig.RedisConfig.Port),
  281. Username: "default",
  282. Password: projectConfig.RedisConfig.Pass,
  283. ReadTimeout: 15 * time.Minute,
  284. })
  285. } else {
  286. projectBackfeedManager.ProjectRedis = that.TrackerRedis
  287. }
  288. go projectBackfeedManager.Do()
  289. that.Lock.Lock()
  290. that.ActiveFeeds[project] = projectBackfeedManager
  291. that.Lock.Unlock()
  292. if outdatedProjectBackfeedManager != nil {
  293. outdatedProjectBackfeedManager.Cancel()
  294. <-outdatedProjectBackfeedManager.Done
  295. log.Printf("updated project: %s", project)
  296. } else {
  297. log.Printf("added project: %s", project)
  298. }
  299. }
  300. that.Lock.Lock()
  301. that.ActiveSlugs = slugProjectMap
  302. that.Lock.Unlock()
  303. // remove feeds for old projects
  304. for project, projectBackfeedManager := range that.ActiveFeeds {
  305. if _, has := projectSlugMap[project]; has {
  306. continue
  307. }
  308. log.Printf("removing project: %s", project)
  309. that.Lock.Lock()
  310. delete(that.ActiveFeeds, project)
  311. that.Lock.Unlock()
  312. projectBackfeedManager.Cancel()
  313. <-projectBackfeedManager.Done
  314. log.Printf("removed project: %s", project)
  315. }
  316. return nil
  317. }
  318. type Splitter struct {
  319. Delimiter []byte
  320. IgnoreEOF bool
  321. }
  322. func (that *Splitter) Split(data []byte, atEOF bool) (int, []byte, error) {
  323. for i := 0; i < len(data); i++ {
  324. if bytes.Equal(data[i:i+len(that.Delimiter)], that.Delimiter) {
  325. return i + len(that.Delimiter), data[:i], nil
  326. }
  327. }
  328. if len(data) == 0 || !atEOF {
  329. return 0, nil, nil
  330. }
  331. if atEOF && that.IgnoreEOF {
  332. return len(data), data, nil
  333. }
  334. return 0, data, io.ErrUnexpectedEOF
  335. }
  336. func GenShardHash(b []byte) (final byte) {
  337. for i, b := range b {
  338. final = (b ^ final ^ byte(i)) + final + byte(i) + final*byte(i)
  339. }
  340. return final
  341. }
  342. func WriteResponse(res http.ResponseWriter, statusCode int, v interface{}) {
  343. res.Header().Set("Content-Type", "application/json")
  344. res.WriteHeader(statusCode)
  345. if statusCode == http.StatusNoContent {
  346. return
  347. }
  348. if err, isError := v.(error); isError {
  349. v = map[string]interface{}{
  350. "error": fmt.Sprintf("%v", err),
  351. "status_code": statusCode,
  352. }
  353. } else {
  354. log.Printf("%#v", v)
  355. v = map[string]interface{}{
  356. "data": v,
  357. "status_code": statusCode,
  358. }
  359. }
  360. json.NewEncoder(res).Encode(v)
  361. }
  362. func (that *GlobalBackfeedManager) GetFeed(slug string) *ProjectBackfeedManager {
  363. that.Lock.RLock()
  364. defer that.Lock.RUnlock()
  365. project, has := that.ActiveSlugs[slug]
  366. if !has {
  367. return nil
  368. }
  369. projectBackfeedManager, has := that.ActiveFeeds[project]
  370. if !has {
  371. return nil
  372. }
  373. return projectBackfeedManager
  374. }
  375. func (that *GlobalBackfeedManager) Handle(res http.ResponseWriter, req *http.Request) {
  376. defer req.Body.Close()
  377. vars := mux.Vars(req)
  378. slug := vars["slug"]
  379. secondaryShard := req.URL.Query().Get("shard")
  380. projectBackfeedManager := that.GetFeed(slug)
  381. if projectBackfeedManager == nil {
  382. WriteResponse(res, http.StatusNotFound, fmt.Errorf("%s", "no such backfeed channel"))
  383. return
  384. }
  385. splitter := &Splitter{
  386. Delimiter: []byte(req.URL.Query().Get("delimiter")),
  387. IgnoreEOF: req.URL.Query().Get("ignoreeof") != "",
  388. }
  389. if len(splitter.Delimiter) == 0 {
  390. splitter.Delimiter = []byte{0x00}
  391. }
  392. scanner := bufio.NewScanner(req.Body)
  393. scanner.Split(splitter.Split)
  394. var err error
  395. statusCode := http.StatusNoContent
  396. n := 0
  397. for scanner.Scan() {
  398. b := scanner.Bytes()
  399. if len(b) == 0 {
  400. continue
  401. }
  402. bcopy := make([]byte, 0, len(b))
  403. copy(bcopy, b)
  404. item := &BackfeedItem{
  405. PrimaryShard: GenShardHash(bcopy),
  406. SecondaryShard: secondaryShard,
  407. Item: bcopy,
  408. }
  409. ok := projectBackfeedManager.PushItem(req.Context(), item)
  410. if !ok {
  411. err = fmt.Errorf("channel closed")
  412. statusCode = http.StatusServiceUnavailable
  413. break
  414. }
  415. n++
  416. }
  417. if err == nil {
  418. err = scanner.Err()
  419. if err != nil {
  420. statusCode = http.StatusBadRequest
  421. }
  422. }
  423. if err != nil {
  424. WriteResponse(res, statusCode, err)
  425. } else {
  426. WriteResponse(res, http.StatusOK, fmt.Sprintf("%d items queued for deduplication", n))
  427. }
  428. return
  429. }
  430. func (that *GlobalBackfeedManager) CancelAllFeeds() {
  431. that.Cancel()
  432. for project, projectBackfeedManager := range that.ActiveFeeds {
  433. log.Printf("waiting for %s channel to shut down...", project)
  434. <-projectBackfeedManager.Done
  435. delete(that.ActiveFeeds, project)
  436. }
  437. }
  438. func main() {
  439. log.SetFlags(log.Flags() | log.Lshortfile)
  440. trackerRedisOptions, err := redis.ParseURL(os.Getenv("REDIS_TRACKER"))
  441. if err != nil {
  442. log.Panicf("invalid REDIS_TRACKER url: %s", err)
  443. }
  444. trackerRedisOptions.ReadTimeout = 15 * time.Minute
  445. trackerRedisClient := redis.NewClient(trackerRedisOptions)
  446. legacyRedisOptions, err := redis.ParseURL(os.Getenv("REDIS_LEGACY"))
  447. if err != nil {
  448. log.Panicf("invalid REDIS_LEGACY url: %s", err)
  449. }
  450. legacyRedisOptions.ReadTimeout = 15 * time.Minute
  451. legacyRedisClient := redis.NewClient(legacyRedisOptions)
  452. backfeedRedisClient := redis.NewClusterClient(&redis.ClusterOptions{
  453. Addrs: strings.Split(os.Getenv("REDIS_BACKFEED_ADDRS"), ","),
  454. Username: os.Getenv("REDIS_BACKFEED_USERNAME"),
  455. Password: os.Getenv("REDIS_BACKFEED_PASSWORD"),
  456. ReadTimeout: 15 * time.Minute,
  457. })
  458. if err := trackerRedisClient.Ping(context.Background()).Err(); err != nil {
  459. log.Panicf("unable to ping tracker redis: %s", err)
  460. }
  461. if err := backfeedRedisClient.Ping(context.Background()).Err(); err != nil {
  462. log.Panicf("unable to ping backfeed redis: %s", err)
  463. }
  464. if err := legacyRedisClient.Ping(context.Background()).Err(); err != nil {
  465. log.Panicf("unable to ping legacy redis: %s", err)
  466. }
  467. globalBackfeedManager := &GlobalBackfeedManager{
  468. ActiveFeeds: map[string]*ProjectBackfeedManager{},
  469. ActiveSlugs: map[string]string{},
  470. TrackerRedis: trackerRedisClient,
  471. BackfeedRedis: backfeedRedisClient,
  472. LegacyRedis: legacyRedisClient,
  473. }
  474. globalBackfeedManager.Context, globalBackfeedManager.Cancel = context.WithCancel(context.Background())
  475. defer globalBackfeedManager.CancelAllFeeds()
  476. err = globalBackfeedManager.RefreshFeeds()
  477. if err != nil {
  478. log.Panicf("unable to set up backfeed projects: %s", err)
  479. }
  480. r := mux.NewRouter()
  481. r.Methods(http.MethodPost).Path("/legacy/{slug}").HandlerFunc(globalBackfeedManager.Handle)
  482. serveErrChan := make(chan error)
  483. go func() {
  484. s := &http.Server{
  485. Addr: os.Getenv("HTTP_ADDR"),
  486. IdleTimeout: 1 * time.Hour,
  487. MaxHeaderBytes: 1 * 1024 * 1024,
  488. Handler: r,
  489. }
  490. serveErrChan <- s.ListenAndServe()
  491. }()
  492. sc := make(chan os.Signal, 1)
  493. signal.Notify(sc, syscall.SIGINT, syscall.SIGTERM, os.Interrupt, os.Kill)
  494. ticker := time.NewTicker(1 * time.Second)
  495. for {
  496. select {
  497. case <-sc:
  498. return
  499. case <-ticker.C:
  500. }
  501. err = globalBackfeedManager.RefreshFeeds()
  502. if err != nil {
  503. log.Printf("unable to refresh backfeed projects: %s", err)
  504. }
  505. }
  506. }