2019-05-02 18:17:27 -04:00
|
|
|
# frozen_string_literal: true
|
|
|
|
|
2016-06-30 10:55:01 -04:00
|
|
|
require "uri"
|
2017-06-13 07:27:05 -04:00
|
|
|
require "mini_mime"
|
2015-06-01 05:13:56 -04:00
|
|
|
require_dependency "file_store/base_store"
|
2014-09-24 16:52:09 -04:00
|
|
|
require_dependency "s3_helper"
|
2014-04-15 07:04:14 -04:00
|
|
|
require_dependency "file_helper"
|
2013-07-31 17:26:34 -04:00
|
|
|
|
2013-11-05 13:04:47 -05:00
|
|
|
module FileStore
|
2013-07-31 17:26:34 -04:00
|
|
|
|
2013-11-05 13:04:47 -05:00
|
|
|
class S3Store < BaseStore
|
2015-05-25 11:59:00 -04:00
|
|
|
TOMBSTONE_PREFIX ||= "tombstone/"
|
|
|
|
|
2018-09-10 03:14:30 -04:00
|
|
|
attr_reader :s3_helper
|
|
|
|
|
2017-07-27 21:20:09 -04:00
|
|
|
def initialize(s3_helper = nil)
|
2018-12-19 00:32:32 -05:00
|
|
|
@s3_helper = s3_helper || S3Helper.new(s3_bucket,
|
|
|
|
Rails.configuration.multisite ? multisite_tombstone_prefix : TOMBSTONE_PREFIX
|
|
|
|
)
|
2014-09-24 16:52:09 -04:00
|
|
|
end
|
2013-07-31 17:26:34 -04:00
|
|
|
|
2015-05-29 12:39:47 -04:00
|
|
|
def store_upload(file, upload, content_type = nil)
|
2016-08-15 04:06:29 -04:00
|
|
|
path = get_path_for_upload(upload)
|
2019-06-05 23:27:24 -04:00
|
|
|
url, upload.etag = store_file(file, path, filename: upload.original_filename, content_type: content_type, cache_locally: true, private: upload.private?)
|
2019-01-04 01:16:22 -05:00
|
|
|
url
|
2013-11-05 13:04:47 -05:00
|
|
|
end
|
2013-07-31 17:26:34 -04:00
|
|
|
|
2016-10-17 13:16:29 -04:00
|
|
|
def store_optimized_image(file, optimized_image, content_type = nil)
|
2016-08-15 04:06:29 -04:00
|
|
|
path = get_path_for_optimized_image(optimized_image)
|
2019-01-04 01:16:22 -05:00
|
|
|
url, optimized_image.etag = store_file(file, path, content_type: content_type)
|
|
|
|
url
|
2016-08-12 05:18:19 -04:00
|
|
|
end
|
|
|
|
|
2015-05-29 12:39:47 -04:00
|
|
|
# options
|
|
|
|
# - filename
|
|
|
|
# - content_type
|
|
|
|
# - cache_locally
|
2017-07-27 21:20:09 -04:00
|
|
|
def store_file(file, path, opts = {})
|
2019-05-02 18:17:27 -04:00
|
|
|
path = path.dup
|
|
|
|
|
2016-10-17 13:16:29 -04:00
|
|
|
filename = opts[:filename].presence || File.basename(path)
|
2015-05-29 12:39:47 -04:00
|
|
|
# cache file locally when needed
|
|
|
|
cache_file(file, File.basename(path)) if opts[:cache_locally]
|
2016-10-17 13:16:29 -04:00
|
|
|
options = {
|
2019-06-05 23:27:24 -04:00
|
|
|
acl: opts[:private] ? "private" : "public-read",
|
2019-08-06 13:55:17 -04:00
|
|
|
cache_control: 'max-age=31556952, public, immutable',
|
2017-06-13 07:27:05 -04:00
|
|
|
content_type: opts[:content_type].presence || MiniMime.lookup_by_filename(filename)&.content_type
|
2016-10-17 13:16:29 -04:00
|
|
|
}
|
2015-05-29 12:39:47 -04:00
|
|
|
# add a "content disposition" header for "attachments"
|
2018-09-09 22:22:45 -04:00
|
|
|
options[:content_disposition] = "attachment; filename=\"#{filename}\"" unless FileHelper.is_supported_image?(filename)
|
2018-11-28 23:11:48 -05:00
|
|
|
|
2018-12-02 23:04:14 -05:00
|
|
|
path.prepend(File.join(upload_path, "/")) if Rails.configuration.multisite
|
2018-11-28 23:11:48 -05:00
|
|
|
|
2019-01-04 01:16:22 -05:00
|
|
|
# if this fails, it will throw an exception
|
|
|
|
path, etag = @s3_helper.upload(file, path, options)
|
|
|
|
|
|
|
|
# return the upload url and etag
|
|
|
|
return File.join(absolute_base_url, path), etag
|
2013-11-05 13:04:47 -05:00
|
|
|
end
|
2013-07-31 17:26:34 -04:00
|
|
|
|
2016-08-14 23:21:24 -04:00
|
|
|
def remove_file(url, path)
|
2015-05-29 12:39:47 -04:00
|
|
|
return unless has_been_uploaded?(url)
|
|
|
|
# copy the removed file to tombstone
|
2016-08-15 04:06:29 -04:00
|
|
|
@s3_helper.remove(path, true)
|
2013-11-05 13:04:47 -05:00
|
|
|
end
|
2013-08-13 16:08:29 -04:00
|
|
|
|
2018-08-07 23:26:05 -04:00
|
|
|
def copy_file(url, source, destination)
|
|
|
|
return unless has_been_uploaded?(url)
|
|
|
|
@s3_helper.copy(source, destination)
|
|
|
|
end
|
|
|
|
|
2013-11-05 13:04:47 -05:00
|
|
|
def has_been_uploaded?(url)
|
2015-05-26 05:47:33 -04:00
|
|
|
return false if url.blank?
|
2016-06-30 10:55:01 -04:00
|
|
|
|
|
|
|
base_hostname = URI.parse(absolute_base_url).hostname
|
|
|
|
return true if url[base_hostname]
|
|
|
|
|
2017-10-06 01:20:01 -04:00
|
|
|
return false if SiteSetting.Upload.s3_cdn_url.blank?
|
|
|
|
cdn_hostname = URI.parse(SiteSetting.Upload.s3_cdn_url || "").hostname
|
2016-06-30 10:55:01 -04:00
|
|
|
cdn_hostname.presence && url[cdn_hostname]
|
2013-11-05 13:04:47 -05:00
|
|
|
end
|
2013-08-13 16:08:29 -04:00
|
|
|
|
2017-10-06 01:20:01 -04:00
|
|
|
def s3_bucket_name
|
|
|
|
@s3_helper.s3_bucket_name
|
|
|
|
end
|
|
|
|
|
2013-11-05 13:04:47 -05:00
|
|
|
def absolute_base_url
|
2017-10-06 01:20:01 -04:00
|
|
|
@absolute_base_url ||= SiteSetting.Upload.absolute_base_url
|
2013-11-05 13:04:47 -05:00
|
|
|
end
|
2013-08-13 16:08:29 -04:00
|
|
|
|
2013-11-05 13:04:47 -05:00
|
|
|
def external?
|
|
|
|
true
|
|
|
|
end
|
2013-08-13 16:08:29 -04:00
|
|
|
|
2013-11-27 16:01:41 -05:00
|
|
|
def purge_tombstone(grace_period)
|
2014-09-24 16:52:09 -04:00
|
|
|
@s3_helper.update_tombstone_lifecycle(grace_period)
|
2013-11-27 16:01:41 -05:00
|
|
|
end
|
|
|
|
|
2018-12-19 00:32:32 -05:00
|
|
|
def multisite_tombstone_prefix
|
|
|
|
File.join("uploads", "tombstone", RailsMultisite::ConnectionManagement.current_db, "/")
|
|
|
|
end
|
|
|
|
|
2019-07-04 11:32:51 -04:00
|
|
|
def download_url(upload)
|
|
|
|
return unless upload
|
|
|
|
"#{upload.short_path}?dl=1"
|
|
|
|
end
|
|
|
|
|
2015-05-25 23:08:31 -04:00
|
|
|
def path_for(upload)
|
2019-05-28 21:00:25 -04:00
|
|
|
url = upload&.url
|
2019-05-29 06:36:09 -04:00
|
|
|
FileStore::LocalStore.new.path_for(upload) if url && url[/^\/[^\/]/]
|
2015-05-25 23:08:31 -04:00
|
|
|
end
|
|
|
|
|
2019-07-04 11:32:51 -04:00
|
|
|
def url_for(upload, force_download: false)
|
|
|
|
if upload.private? || force_download
|
|
|
|
opts = { expires_in: S3Helper::DOWNLOAD_URL_EXPIRES_AFTER_SECONDS }
|
2019-08-07 11:00:43 -04:00
|
|
|
|
|
|
|
if force_download
|
|
|
|
opts[:response_content_disposition] = ActionDispatch::Http::ContentDisposition.format(
|
|
|
|
disposition: "attachment", filename: upload.original_filename
|
|
|
|
)
|
|
|
|
end
|
2019-07-04 11:32:51 -04:00
|
|
|
|
2019-06-05 23:27:24 -04:00
|
|
|
obj = @s3_helper.object(get_upload_key(upload))
|
2019-07-04 11:32:51 -04:00
|
|
|
url = obj.presigned_url(:get, opts)
|
2019-06-05 23:27:24 -04:00
|
|
|
else
|
|
|
|
url = upload.url
|
|
|
|
end
|
|
|
|
|
|
|
|
url
|
|
|
|
end
|
|
|
|
|
2015-05-26 22:02:57 -04:00
|
|
|
def cdn_url(url)
|
2017-10-06 01:20:01 -04:00
|
|
|
return url if SiteSetting.Upload.s3_cdn_url.blank?
|
2016-06-30 10:55:01 -04:00
|
|
|
schema = url[/^(https?:)?\/\//, 1]
|
2018-07-06 18:15:28 -04:00
|
|
|
folder = @s3_helper.s3_bucket_folder_path.nil? ? "" : "#{@s3_helper.s3_bucket_folder_path}/"
|
2018-11-28 23:11:48 -05:00
|
|
|
url.sub(File.join("#{schema}#{absolute_base_url}", folder), File.join(SiteSetting.Upload.s3_cdn_url, "/"))
|
2015-05-26 22:02:57 -04:00
|
|
|
end
|
|
|
|
|
2015-05-29 12:39:47 -04:00
|
|
|
def cache_avatar(avatar, user_id)
|
|
|
|
source = avatar.url.sub(absolute_base_url + "/", "")
|
|
|
|
destination = avatar_template(avatar, user_id).sub(absolute_base_url + "/", "")
|
|
|
|
@s3_helper.copy(source, destination)
|
|
|
|
end
|
2013-11-27 16:01:41 -05:00
|
|
|
|
2015-05-29 12:39:47 -04:00
|
|
|
def avatar_template(avatar, user_id)
|
|
|
|
UserAvatar.external_avatar_url(user_id, avatar.upload_id, avatar.width)
|
|
|
|
end
|
2013-07-31 17:26:34 -04:00
|
|
|
|
2016-08-19 02:08:04 -04:00
|
|
|
def s3_bucket
|
2017-10-06 01:20:01 -04:00
|
|
|
raise Discourse::SiteSettingMissing.new("s3_upload_bucket") if SiteSetting.Upload.s3_upload_bucket.blank?
|
|
|
|
SiteSetting.Upload.s3_upload_bucket.downcase
|
2015-05-29 12:39:47 -04:00
|
|
|
end
|
2018-11-26 14:24:51 -05:00
|
|
|
|
2019-02-14 14:04:35 -05:00
|
|
|
def list_missing_uploads(skip_optimized: false)
|
2019-01-31 23:40:48 -05:00
|
|
|
if SiteSetting.enable_s3_inventory
|
|
|
|
require 's3_inventory'
|
2019-02-19 11:24:35 -05:00
|
|
|
S3Inventory.new(s3_helper, :upload).backfill_etags_and_list_missing
|
|
|
|
S3Inventory.new(s3_helper, :optimized).backfill_etags_and_list_missing unless skip_optimized
|
2019-01-31 23:40:48 -05:00
|
|
|
else
|
2019-03-13 05:39:07 -04:00
|
|
|
list_missing(Upload.by_users, "original/")
|
2019-01-31 23:40:48 -05:00
|
|
|
list_missing(OptimizedImage, "optimized/") unless skip_optimized
|
|
|
|
end
|
2018-11-26 14:24:51 -05:00
|
|
|
end
|
|
|
|
|
2019-06-05 23:27:24 -04:00
|
|
|
def update_upload_ACL(upload)
|
|
|
|
private_uploads = SiteSetting.prevent_anons_from_downloading_files
|
|
|
|
key = get_upload_key(upload)
|
|
|
|
|
|
|
|
begin
|
|
|
|
@s3_helper.object(key).acl.put(acl: private_uploads ? "private" : "public-read")
|
|
|
|
rescue Aws::S3::Errors::NoSuchKey
|
|
|
|
Rails.logger.warn("Could not update ACL on upload with key: '#{key}'. Upload is missing.")
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2019-07-01 14:38:36 -04:00
|
|
|
def download_file(upload, destination_path)
|
|
|
|
@s3_helper.download_file(get_upload_key(upload), destination_path)
|
|
|
|
end
|
|
|
|
|
2018-11-26 14:24:51 -05:00
|
|
|
private
|
|
|
|
|
2019-06-05 23:27:24 -04:00
|
|
|
def get_upload_key(upload)
|
|
|
|
if Rails.configuration.multisite
|
|
|
|
File.join(upload_path, "/", get_path_for_upload(upload))
|
|
|
|
else
|
|
|
|
get_path_for_upload(upload)
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2018-11-26 14:24:51 -05:00
|
|
|
def list_missing(model, prefix)
|
|
|
|
connection = ActiveRecord::Base.connection.raw_connection
|
|
|
|
connection.exec('CREATE TEMP TABLE verified_ids(val integer PRIMARY KEY)')
|
|
|
|
marker = nil
|
|
|
|
files = @s3_helper.list(prefix, marker)
|
|
|
|
|
|
|
|
while files.count > 0 do
|
|
|
|
verified_ids = []
|
|
|
|
|
|
|
|
files.each do |f|
|
2019-10-21 06:32:27 -04:00
|
|
|
id = model.where("url LIKE '%#{f.key}' AND etag = '#{f.etag}'").pluck_first(:id)
|
2018-11-26 14:24:51 -05:00
|
|
|
verified_ids << id if id.present?
|
|
|
|
marker = f.key
|
|
|
|
end
|
|
|
|
|
|
|
|
verified_id_clause = verified_ids.map { |id| "('#{PG::Connection.escape_string(id.to_s)}')" }.join(",")
|
|
|
|
connection.exec("INSERT INTO verified_ids VALUES #{verified_id_clause}")
|
|
|
|
files = @s3_helper.list(prefix, marker)
|
|
|
|
end
|
|
|
|
|
2019-01-31 23:40:48 -05:00
|
|
|
missing_uploads = model.joins('LEFT JOIN verified_ids ON verified_ids.val = id').where("verified_ids.val IS NULL")
|
2018-11-26 14:24:51 -05:00
|
|
|
missing_count = missing_uploads.count
|
|
|
|
|
|
|
|
if missing_count > 0
|
|
|
|
missing_uploads.find_each do |upload|
|
|
|
|
puts upload.url
|
|
|
|
end
|
|
|
|
|
|
|
|
puts "#{missing_count} of #{model.count} #{model.name.underscore.pluralize} are missing"
|
|
|
|
end
|
|
|
|
ensure
|
2018-11-26 14:45:29 -05:00
|
|
|
connection.exec('DROP TABLE verified_ids') unless connection.nil?
|
2018-11-26 14:24:51 -05:00
|
|
|
end
|
2013-07-31 17:26:34 -04:00
|
|
|
end
|
|
|
|
end
|