about summary refs log tree commit diff
path: root/streaming
diff options
context:
space:
mode:
authorYamagishi Kazutoshi <ykzts@desire.sh>2017-05-21 00:31:47 +0900
committerEugen Rochko <eugen@zeonfederated.com>2017-05-20 17:31:47 +0200
commit2e112e240666b62b8c3d4fa201fb24b841f6c92b (patch)
treef95f2eb5ae8156c10bf16894919d0ba5e3f34096 /streaming
parent812fe90ecaef58f1dbe16c3fdfea79e14e7bbc9d (diff)
Improve eslint rules (#3147)
* Add semi to ESLint rules

* Add padded-blocks to ESLint rules

* Add comma-dangle to ESLint rules

* add config/webpack and storyboard

* add streaming/

* yarn test:lint -- --fix
Diffstat (limited to 'streaming')
-rw-r--r--streaming/index.js356
1 files changed, 178 insertions, 178 deletions
diff --git a/streaming/index.js b/streaming/index.js
index 7f30b4e88..cd1963121 100644
--- a/streaming/index.js
+++ b/streaming/index.js
@@ -1,56 +1,56 @@
 import os from 'os';
 import cluster from 'cluster';
-import dotenv from 'dotenv'
-import express from 'express'
-import http from 'http'
-import redis from 'redis'
-import pg from 'pg'
-import log from 'npmlog'
-import url from 'url'
-import WebSocket from 'uws'
-import uuid from 'uuid'
-
-const env = process.env.NODE_ENV || 'development'
+import dotenv from 'dotenv';
+import express from 'express';
+import http from 'http';
+import redis from 'redis';
+import pg from 'pg';
+import log from 'npmlog';
+import url from 'url';
+import WebSocket from 'uws';
+import uuid from 'uuid';
+
+const env = process.env.NODE_ENV || 'development';
 
 dotenv.config({
-  path: env === 'production' ? '.env.production' : '.env'
-})
+  path: env === 'production' ? '.env.production' : '.env',
+});
 
 const dbUrlToConfig = (dbUrl) => {
   if (!dbUrl) {
-    return {}
+    return {};
   }
 
-  const params = url.parse(dbUrl)
-  const config = {}
+  const params = url.parse(dbUrl);
+  const config = {};
 
   if (params.auth) {
-    [config.user, config.password] = params.auth.split(':')
+    [config.user, config.password] = params.auth.split(':');
   }
 
   if (params.hostname) {
-    config.host = params.hostname
+    config.host = params.hostname;
   }
 
   if (params.port) {
-    config.port = params.port
+    config.port = params.port;
   }
 
   if (params.pathname) {
-    config.database = params.pathname.split('/')[1]
+    config.database = params.pathname.split('/')[1];
   }
 
-  const ssl = params.query && params.query.ssl
+  const ssl = params.query && params.query.ssl;
   if (ssl) {
-    config.ssl = ssl === 'true' || ssl === '1'
+    config.ssl = ssl === 'true' || ssl === '1';
   }
 
-  return config
-}
+  return config;
+};
 
 if (cluster.isMaster) {
   // Cluster master
-  const core = +process.env.STREAMING_CLUSTER_NUM || (env === 'development' ? 1 : Math.max(os.cpus().length - 1, 1))
+  const core = +process.env.STREAMING_CLUSTER_NUM || (env === 'development' ? 1 : Math.max(os.cpus().length - 1, 1));
 
   const fork = () => {
     const worker = cluster.fork();
@@ -63,14 +63,14 @@ if (cluster.isMaster) {
 
   for (let i = 0; i < core; i++) fork();
 
-  log.info(`Starting streaming API server master with ${core} workers`)
+  log.info(`Starting streaming API server master with ${core} workers`);
 } else {
   // Cluster worker
   const pgConfigs = {
     development: {
       database: 'mastodon_development',
       host:     '/var/run/postgresql',
-      max:      10
+      max:      10,
     },
 
     production: {
@@ -79,315 +79,315 @@ if (cluster.isMaster) {
       database: process.env.DB_NAME || 'mastodon_production',
       host:     process.env.DB_HOST || 'localhost',
       port:     process.env.DB_PORT || 5432,
-      max:      10
-    }
-  }
+      max:      10,
+    },
+  };
 
-  const app    = express()
-  const pgPool = new pg.Pool(Object.assign(pgConfigs[env], dbUrlToConfig(process.env.DATABASE_URL)))
-  const server = http.createServer(app)
-  const wss    = new WebSocket.Server({ server })
-  const redisNamespace = process.env.REDIS_NAMESPACE || null
+  const app    = express();
+  const pgPool = new pg.Pool(Object.assign(pgConfigs[env], dbUrlToConfig(process.env.DATABASE_URL)));
+  const server = http.createServer(app);
+  const wss    = new WebSocket.Server({ server });
+  const redisNamespace = process.env.REDIS_NAMESPACE || null;
 
   const redisParams = {
     host:     process.env.REDIS_HOST     || '127.0.0.1',
     port:     process.env.REDIS_PORT     || 6379,
     db:       process.env.REDIS_DB       || 0,
     password: process.env.REDIS_PASSWORD,
-    url:      process.env.REDIS_URL      || null
-  }
+    url:      process.env.REDIS_URL      || null,
+  };
 
   if (redisNamespace) {
-    redisParams.namespace = redisNamespace
+    redisParams.namespace = redisNamespace;
   }
-  const redisPrefix = redisNamespace ? `${redisNamespace}:` : ''
+  const redisPrefix = redisNamespace ? `${redisNamespace}:` : '';
 
-  const redisClient = redis.createClient(redisParams)
+  const redisClient = redis.createClient(redisParams);
 
-  const subs = {}
+  const subs = {};
 
   redisClient.on('pmessage', (_, channel, message) => {
-    const callbacks = subs[channel]
+    const callbacks = subs[channel];
 
-    log.silly(`New message on channel ${channel}`)
+    log.silly(`New message on channel ${channel}`);
 
     if (!callbacks) {
-      return
+      return;
     }
-    callbacks.forEach(callback => callback(message))
-  })
+    callbacks.forEach(callback => callback(message));
+  });
 
-  redisClient.psubscribe(`${redisPrefix}timeline:*`)
+  redisClient.psubscribe(`${redisPrefix}timeline:*`);
 
   const subscribe = (channel, callback) => {
-    log.silly(`Adding listener for ${channel}`)
-    subs[channel] = subs[channel] || []
-    subs[channel].push(callback)
-  }
+    log.silly(`Adding listener for ${channel}`);
+    subs[channel] = subs[channel] || [];
+    subs[channel].push(callback);
+  };
 
   const unsubscribe = (channel, callback) => {
-    log.silly(`Removing listener for ${channel}`)
-    subs[channel] = subs[channel].filter(item => item !== callback)
-  }
+    log.silly(`Removing listener for ${channel}`);
+    subs[channel] = subs[channel].filter(item => item !== callback);
+  };
 
   const allowCrossDomain = (req, res, next) => {
-    res.header('Access-Control-Allow-Origin', '*')
-    res.header('Access-Control-Allow-Headers', 'Authorization, Accept, Cache-Control')
-    res.header('Access-Control-Allow-Methods', 'GET, OPTIONS')
+    res.header('Access-Control-Allow-Origin', '*');
+    res.header('Access-Control-Allow-Headers', 'Authorization, Accept, Cache-Control');
+    res.header('Access-Control-Allow-Methods', 'GET, OPTIONS');
 
-    next()
-  }
+    next();
+  };
 
   const setRequestId = (req, res, next) => {
-    req.requestId = uuid.v4()
-    res.header('X-Request-Id', req.requestId)
+    req.requestId = uuid.v4();
+    res.header('X-Request-Id', req.requestId);
 
-    next()
-  }
+    next();
+  };
 
   const accountFromToken = (token, req, next) => {
     pgPool.connect((err, client, done) => {
       if (err) {
-        next(err)
-        return
+        next(err);
+        return;
       }
 
       client.query('SELECT oauth_access_tokens.resource_owner_id, users.account_id FROM oauth_access_tokens INNER JOIN users ON oauth_access_tokens.resource_owner_id = users.id WHERE oauth_access_tokens.token = $1 LIMIT 1', [token], (err, result) => {
-        done()
+        done();
 
         if (err) {
-          next(err)
-          return
+          next(err);
+          return;
         }
 
         if (result.rows.length === 0) {
-          err = new Error('Invalid access token')
-          err.statusCode = 401
+          err = new Error('Invalid access token');
+          err.statusCode = 401;
 
-          next(err)
-          return
+          next(err);
+          return;
         }
 
-        req.accountId = result.rows[0].account_id
+        req.accountId = result.rows[0].account_id;
 
-        next()
-      })
-    })
-  }
+        next();
+      });
+    });
+  };
 
   const authenticationMiddleware = (req, res, next) => {
     if (req.method === 'OPTIONS') {
-      next()
-      return
+      next();
+      return;
     }
 
-    const authorization = req.get('Authorization')
+    const authorization = req.get('Authorization');
 
     if (!authorization) {
-      const err = new Error('Missing access token')
-      err.statusCode = 401
+      const err = new Error('Missing access token');
+      err.statusCode = 401;
 
-      next(err)
-      return
+      next(err);
+      return;
     }
 
-    const token = authorization.replace(/^Bearer /, '')
+    const token = authorization.replace(/^Bearer /, '');
 
-    accountFromToken(token, req, next)
-  }
+    accountFromToken(token, req, next);
+  };
 
   const errorMiddleware = (err, req, res, next) => {
-    log.error(req.requestId, err)
-    res.writeHead(err.statusCode || 500, { 'Content-Type': 'application/json' })
-    res.end(JSON.stringify({ error: err.statusCode ? `${err}` : 'An unexpected error occurred' }))
-  }
+    log.error(req.requestId, err);
+    res.writeHead(err.statusCode || 500, { 'Content-Type': 'application/json' });
+    res.end(JSON.stringify({ error: err.statusCode ? `${err}` : 'An unexpected error occurred' }));
+  };
 
   const placeholders = (arr, shift = 0) => arr.map((_, i) => `$${i + 1 + shift}`).join(', ');
 
   const streamFrom = (id, req, output, attachCloseHandler, needsFiltering = false) => {
-    log.verbose(req.requestId, `Starting stream from ${id} for ${req.accountId}`)
+    log.verbose(req.requestId, `Starting stream from ${id} for ${req.accountId}`);
 
     const listener = message => {
-      const { event, payload, queued_at } = JSON.parse(message)
+      const { event, payload, queued_at } = JSON.parse(message);
 
       const transmit = () => {
-        const now   = new Date().getTime()
+        const now   = new Date().getTime();
         const delta = now - queued_at;
 
-        log.silly(req.requestId, `Transmitting for ${req.accountId}: ${event} ${payload} Delay: ${delta}ms`)
-        output(event, payload)
-      }
+        log.silly(req.requestId, `Transmitting for ${req.accountId}: ${event} ${payload} Delay: ${delta}ms`);
+        output(event, payload);
+      };
 
       // Only messages that may require filtering are statuses, since notifications
       // are already personalized and deletes do not matter
       if (needsFiltering && event === 'update') {
         pgPool.connect((err, client, done) => {
           if (err) {
-            log.error(err)
-            return
+            log.error(err);
+            return;
           }
 
-          const unpackedPayload  = JSON.parse(payload)
-          const targetAccountIds = [unpackedPayload.account.id].concat(unpackedPayload.mentions.map(item => item.id)).concat(unpackedPayload.reblog ? [unpackedPayload.reblog.account.id] : [])
-          const accountDomain    = unpackedPayload.account.acct.split('@')[1]
+          const unpackedPayload  = JSON.parse(payload);
+          const targetAccountIds = [unpackedPayload.account.id].concat(unpackedPayload.mentions.map(item => item.id)).concat(unpackedPayload.reblog ? [unpackedPayload.reblog.account.id] : []);
+          const accountDomain    = unpackedPayload.account.acct.split('@')[1];
 
           const queries = [
             client.query(`SELECT 1 FROM blocks WHERE account_id = $1 AND target_account_id IN (${placeholders(targetAccountIds, 1)}) UNION SELECT 1 FROM mutes WHERE account_id = $1 AND target_account_id IN (${placeholders(targetAccountIds, 1)})`, [req.accountId].concat(targetAccountIds)),
-          ]
+          ];
 
           if (accountDomain) {
-            queries.push(client.query('SELECT 1 FROM account_domain_blocks WHERE account_id = $1 AND domain = $2', [req.accountId, accountDomain]))
+            queries.push(client.query('SELECT 1 FROM account_domain_blocks WHERE account_id = $1 AND domain = $2', [req.accountId, accountDomain]));
           }
 
           Promise.all(queries).then(values => {
-            done()
+            done();
 
             if (values[0].rows.length > 0 || (values.length > 1 && values[1].rows.length > 0)) {
-              return
+              return;
             }
 
-            transmit()
+            transmit();
           }).catch(err => {
-            log.error(err)
-          })
-        })
+            log.error(err);
+          });
+        });
       } else {
-        transmit()
+        transmit();
       }
-    }
+    };
 
-    subscribe(`${redisPrefix}${id}`, listener)
-    attachCloseHandler(`${redisPrefix}${id}`, listener)
-  }
+    subscribe(`${redisPrefix}${id}`, listener);
+    attachCloseHandler(`${redisPrefix}${id}`, listener);
+  };
 
   // Setup stream output to HTTP
   const streamToHttp = (req, res) => {
-    res.setHeader('Content-Type', 'text/event-stream')
-    res.setHeader('Transfer-Encoding', 'chunked')
+    res.setHeader('Content-Type', 'text/event-stream');
+    res.setHeader('Transfer-Encoding', 'chunked');
 
-    const heartbeat = setInterval(() => res.write(':thump\n'), 15000)
+    const heartbeat = setInterval(() => res.write(':thump\n'), 15000);
 
     req.on('close', () => {
-      log.verbose(req.requestId, `Ending stream for ${req.accountId}`)
-      clearInterval(heartbeat)
-    })
+      log.verbose(req.requestId, `Ending stream for ${req.accountId}`);
+      clearInterval(heartbeat);
+    });
 
     return (event, payload) => {
-      res.write(`event: ${event}\n`)
-      res.write(`data: ${payload}\n\n`)
-    }
-  }
+      res.write(`event: ${event}\n`);
+      res.write(`data: ${payload}\n\n`);
+    };
+  };
 
   // Setup stream end for HTTP
   const streamHttpEnd = req => (id, listener) => {
     req.on('close', () => {
-      unsubscribe(id, listener)
-    })
-  }
+      unsubscribe(id, listener);
+    });
+  };
 
   // Setup stream output to WebSockets
   const streamToWs = (req, ws) => {
     const heartbeat = setInterval(() => {
       // TODO: Can't add multiple listeners, due to the limitation of uws.
       if (ws.readyState !== ws.OPEN) {
-        log.verbose(req.requestId, `Ending stream for ${req.accountId}`)
-        clearInterval(heartbeat)
-        return
+        log.verbose(req.requestId, `Ending stream for ${req.accountId}`);
+        clearInterval(heartbeat);
+        return;
       }
 
-      ws.ping()
-    }, 15000)
+      ws.ping();
+    }, 15000);
 
     return (event, payload) => {
       if (ws.readyState !== ws.OPEN) {
-        log.error(req.requestId, 'Tried writing to closed socket')
-        return
+        log.error(req.requestId, 'Tried writing to closed socket');
+        return;
       }
 
-      ws.send(JSON.stringify({ event, payload }))
-    }
-  }
+      ws.send(JSON.stringify({ event, payload }));
+    };
+  };
 
   // Setup stream end for WebSockets
   const streamWsEnd = ws => (id, listener) => {
     ws.on('close', () => {
-      unsubscribe(id, listener)
-    })
+      unsubscribe(id, listener);
+    });
 
     ws.on('error', e => {
-      unsubscribe(id, listener)
-    })
-  }
+      unsubscribe(id, listener);
+    });
+  };
 
-  app.use(setRequestId)
-  app.use(allowCrossDomain)
-  app.use(authenticationMiddleware)
-  app.use(errorMiddleware)
+  app.use(setRequestId);
+  app.use(allowCrossDomain);
+  app.use(authenticationMiddleware);
+  app.use(errorMiddleware);
 
   app.get('/api/v1/streaming/user', (req, res) => {
-    streamFrom(`timeline:${req.accountId}`, req, streamToHttp(req, res), streamHttpEnd(req))
-  })
+    streamFrom(`timeline:${req.accountId}`, req, streamToHttp(req, res), streamHttpEnd(req));
+  });
 
   app.get('/api/v1/streaming/public', (req, res) => {
-    streamFrom('timeline:public', req, streamToHttp(req, res), streamHttpEnd(req), true)
-  })
+    streamFrom('timeline:public', req, streamToHttp(req, res), streamHttpEnd(req), true);
+  });
 
   app.get('/api/v1/streaming/public/local', (req, res) => {
-    streamFrom('timeline:public:local', req, streamToHttp(req, res), streamHttpEnd(req), true)
-  })
+    streamFrom('timeline:public:local', req, streamToHttp(req, res), streamHttpEnd(req), true);
+  });
 
   app.get('/api/v1/streaming/hashtag', (req, res) => {
-    streamFrom(`timeline:hashtag:${req.query.tag}`, req, streamToHttp(req, res), streamHttpEnd(req), true)
-  })
+    streamFrom(`timeline:hashtag:${req.query.tag}`, req, streamToHttp(req, res), streamHttpEnd(req), true);
+  });
 
   app.get('/api/v1/streaming/hashtag/local', (req, res) => {
-    streamFrom(`timeline:hashtag:${req.query.tag}:local`, req, streamToHttp(req, res), streamHttpEnd(req), true)
-  })
+    streamFrom(`timeline:hashtag:${req.query.tag}:local`, req, streamToHttp(req, res), streamHttpEnd(req), true);
+  });
 
   wss.on('connection', ws => {
-    const location = url.parse(ws.upgradeReq.url, true)
-    const token    = location.query.access_token
-    const req      = { requestId: uuid.v4() }
+    const location = url.parse(ws.upgradeReq.url, true);
+    const token    = location.query.access_token;
+    const req      = { requestId: uuid.v4() };
 
     accountFromToken(token, req, err => {
       if (err) {
-        log.error(req.requestId, err)
-        ws.close()
-        return
+        log.error(req.requestId, err);
+        ws.close();
+        return;
       }
 
       switch(location.query.stream) {
       case 'user':
-        streamFrom(`timeline:${req.accountId}`, req, streamToWs(req, ws), streamWsEnd(ws))
+        streamFrom(`timeline:${req.accountId}`, req, streamToWs(req, ws), streamWsEnd(ws));
         break;
       case 'public':
-        streamFrom('timeline:public', req, streamToWs(req, ws), streamWsEnd(ws), true)
+        streamFrom('timeline:public', req, streamToWs(req, ws), streamWsEnd(ws), true);
         break;
       case 'public:local':
-        streamFrom('timeline:public:local', req, streamToWs(req, ws), streamWsEnd(ws), true)
+        streamFrom('timeline:public:local', req, streamToWs(req, ws), streamWsEnd(ws), true);
         break;
       case 'hashtag':
-        streamFrom(`timeline:hashtag:${location.query.tag}`, req, streamToWs(req, ws), streamWsEnd(ws), true)
+        streamFrom(`timeline:hashtag:${location.query.tag}`, req, streamToWs(req, ws), streamWsEnd(ws), true);
         break;
       case 'hashtag:local':
-        streamFrom(`timeline:hashtag:${location.query.tag}:local`, req, streamToWs(req, ws), streamWsEnd(ws), true)
+        streamFrom(`timeline:hashtag:${location.query.tag}:local`, req, streamToWs(req, ws), streamWsEnd(ws), true);
         break;
       default:
-        ws.close()
+        ws.close();
       }
-    })
-  })
+    });
+  });
 
   server.listen(process.env.PORT || 4000, () => {
-    log.level = process.env.LOG_LEVEL || 'verbose'
-    log.info(`Starting streaming API server worker on ${server.address().address}:${server.address().port}`)
-  })
+    log.level = process.env.LOG_LEVEL || 'verbose';
+    log.info(`Starting streaming API server worker on ${server.address().address}:${server.address().port}`);
+  });
 
-  process.on('SIGINT', exit)
-  process.on('SIGTERM', exit)
-  process.on('exit', exit)
+  process.on('SIGINT', exit);
+  process.on('SIGTERM', exit);
+  process.on('exit', exit);
 
   function exit() {
-    server.close()
+    server.close();
   }
 }