chore: Move to prettier standard double quotes (#1309)

This commit is contained in:
Tom Moor
2020-06-20 13:59:15 -07:00
committed by GitHub
parent 2a3b9e2104
commit f43deb7940
444 changed files with 5988 additions and 5977 deletions

View File

@ -1,11 +1,11 @@
// @flow
import crypto from 'crypto';
import addHours from 'date-fns/add_hours';
import format from 'date-fns/format';
import AWS from 'aws-sdk';
import invariant from 'invariant';
import fetch from 'isomorphic-fetch';
import * as Sentry from '@sentry/node';
import crypto from "crypto";
import addHours from "date-fns/add_hours";
import format from "date-fns/format";
import AWS from "aws-sdk";
import invariant from "invariant";
import fetch from "isomorphic-fetch";
import * as Sentry from "@sentry/node";
const AWS_SECRET_ACCESS_KEY = process.env.AWS_SECRET_ACCESS_KEY;
const AWS_ACCESS_KEY_ID = process.env.AWS_ACCESS_KEY_ID;
@ -17,24 +17,24 @@ const s3 = new AWS.S3({
accessKeyId: process.env.AWS_ACCESS_KEY_ID,
secretAccessKey: process.env.AWS_SECRET_ACCESS_KEY,
endpoint: new AWS.Endpoint(process.env.AWS_S3_UPLOAD_BUCKET_URL),
signatureVersion: 'v4',
signatureVersion: "v4",
});
const hmac = (key: string, message: string, encoding: any) => {
return crypto
.createHmac('sha256', key)
.update(message, 'utf8')
.createHmac("sha256", key)
.update(message, "utf8")
.digest(encoding);
};
export const makeCredential = () => {
const credential =
AWS_ACCESS_KEY_ID +
'/' +
format(new Date(), 'YYYYMMDD') +
'/' +
"/" +
format(new Date(), "YYYYMMDD") +
"/" +
AWS_REGION +
'/s3/aws4_request';
"/s3/aws4_request";
return credential;
};
@ -47,31 +47,31 @@ export const makePolicy = (
const policy = {
conditions: [
{ bucket: process.env.AWS_S3_UPLOAD_BUCKET_NAME },
['starts-with', '$key', ''],
["starts-with", "$key", ""],
{ acl },
['content-length-range', 0, +process.env.AWS_S3_UPLOAD_MAX_SIZE],
['starts-with', '$Content-Type', 'image'],
['starts-with', '$Cache-Control', ''],
{ 'x-amz-algorithm': 'AWS4-HMAC-SHA256' },
{ 'x-amz-credential': credential },
{ 'x-amz-date': longDate },
["content-length-range", 0, +process.env.AWS_S3_UPLOAD_MAX_SIZE],
["starts-with", "$Content-Type", "image"],
["starts-with", "$Cache-Control", ""],
{ "x-amz-algorithm": "AWS4-HMAC-SHA256" },
{ "x-amz-credential": credential },
{ "x-amz-date": longDate },
],
expiration: format(tomorrow, 'YYYY-MM-DDTHH:mm:ss\\Z'),
expiration: format(tomorrow, "YYYY-MM-DDTHH:mm:ss\\Z"),
};
return new Buffer(JSON.stringify(policy)).toString('base64');
return new Buffer(JSON.stringify(policy)).toString("base64");
};
export const getSignature = (policy: any) => {
const kDate = hmac(
'AWS4' + AWS_SECRET_ACCESS_KEY,
format(new Date(), 'YYYYMMDD')
"AWS4" + AWS_SECRET_ACCESS_KEY,
format(new Date(), "YYYYMMDD")
);
const kRegion = hmac(kDate, AWS_REGION);
const kService = hmac(kRegion, 's3');
const kCredentials = hmac(kService, 'aws4_request');
const kService = hmac(kRegion, "s3");
const kCredentials = hmac(kService, "aws4_request");
const signature = hmac(kCredentials, policy, 'hex');
const signature = hmac(kCredentials, policy, "hex");
return signature;
};
@ -80,11 +80,11 @@ export const publicS3Endpoint = (isServerUpload?: boolean) => {
// for access outside of docker containers in local development
const isDocker = process.env.AWS_S3_UPLOAD_BUCKET_URL.match(/http:\/\/s3:/);
const host = process.env.AWS_S3_UPLOAD_BUCKET_URL.replace(
's3:',
'localhost:'
).replace(/\/$/, '');
"s3:",
"localhost:"
).replace(/\/$/, "");
return `${host}/${isServerUpload && isDocker ? 's3/' : ''}${
return `${host}/${isServerUpload && isDocker ? "s3/" : ""}${
process.env.AWS_S3_UPLOAD_BUCKET_NAME
}`;
};
@ -94,7 +94,7 @@ export const uploadToS3FromUrl = async (
key: string,
acl: string
) => {
invariant(AWS_S3_UPLOAD_BUCKET_NAME, 'AWS_S3_UPLOAD_BUCKET_NAME not set');
invariant(AWS_S3_UPLOAD_BUCKET_NAME, "AWS_S3_UPLOAD_BUCKET_NAME not set");
try {
// $FlowIssue https://github.com/facebook/flow/issues/2171
@ -105,9 +105,9 @@ export const uploadToS3FromUrl = async (
ACL: acl,
Bucket: process.env.AWS_S3_UPLOAD_BUCKET_NAME,
Key: key,
ContentType: res.headers['content-type'],
ContentLength: res.headers['content-length'],
ServerSideEncryption: 'AES256',
ContentType: res.headers["content-type"],
ContentLength: res.headers["content-length"],
ServerSideEncryption: "AES256",
Body: buffer,
})
.promise();
@ -133,7 +133,7 @@ export const deleteFromS3 = (key: string) => {
};
export const getSignedImageUrl = async (key: string) => {
invariant(AWS_S3_UPLOAD_BUCKET_NAME, 'AWS_S3_UPLOAD_BUCKET_NAME not set');
invariant(AWS_S3_UPLOAD_BUCKET_NAME, "AWS_S3_UPLOAD_BUCKET_NAME not set");
const isDocker = process.env.AWS_S3_UPLOAD_BUCKET_URL.match(/http:\/\/s3:/);
const params = {
@ -144,7 +144,7 @@ export const getSignedImageUrl = async (key: string) => {
return isDocker
? `${publicS3Endpoint()}/${key}`
: s3.getSignedUrl('getObject', params);
: s3.getSignedUrl("getObject", params);
};
export const getImageByKey = async (key: string) => {