FIX: don't double request when downloading a file
This commit is contained in:
parent
b731d5d9b5
commit
0559a4736a
|
@ -94,6 +94,8 @@ class StaticController < ApplicationController
|
||||||
redirect_to destination
|
redirect_to destination
|
||||||
end
|
end
|
||||||
|
|
||||||
|
FAVICON ||= -"favicon"
|
||||||
|
|
||||||
# We need to be able to draw our favicon on a canvas
|
# We need to be able to draw our favicon on a canvas
|
||||||
# and pull it off the canvas into a data uri
|
# and pull it off the canvas into a data uri
|
||||||
# This can work by ensuring people set all the right CORS
|
# This can work by ensuring people set all the right CORS
|
||||||
|
@ -101,16 +103,16 @@ class StaticController < ApplicationController
|
||||||
# instead we cache the favicon in redis and serve it out real quick with
|
# instead we cache the favicon in redis and serve it out real quick with
|
||||||
# a huge expiry, we also cache these assets in nginx so it bypassed if needed
|
# a huge expiry, we also cache these assets in nginx so it bypassed if needed
|
||||||
def favicon
|
def favicon
|
||||||
|
|
||||||
hijack do
|
hijack do
|
||||||
data = DistributedMemoizer.memoize('favicon' + SiteSetting.favicon_url, 60 * 30) do
|
data = DistributedMemoizer.memoize(FAVICON + SiteSetting.favicon_url, 60 * 30) do
|
||||||
begin
|
begin
|
||||||
file = FileHelper.download(
|
file = FileHelper.download(
|
||||||
SiteSetting.favicon_url,
|
SiteSetting.favicon_url,
|
||||||
max_file_size: 50.kilobytes,
|
max_file_size: 50.kilobytes,
|
||||||
tmp_file_name: "favicon.png",
|
tmp_file_name: FAVICON,
|
||||||
follow_redirect: true
|
follow_redirect: true
|
||||||
)
|
)
|
||||||
|
file ||= Tempfile.new([FAVICON, ".png"])
|
||||||
data = file.read
|
data = file.read
|
||||||
file.unlink
|
file.unlink
|
||||||
data
|
data
|
||||||
|
|
|
@ -25,57 +25,55 @@ class FileHelper
|
||||||
follow_redirect: false,
|
follow_redirect: false,
|
||||||
read_timeout: 5,
|
read_timeout: 5,
|
||||||
skip_rate_limit: false,
|
skip_rate_limit: false,
|
||||||
verbose: nil)
|
verbose: false)
|
||||||
|
|
||||||
# verbose logging is default while debugging onebox
|
|
||||||
verbose = verbose.nil? ? true : verbose
|
|
||||||
|
|
||||||
url = "https:" + url if url.start_with?("//")
|
url = "https:" + url if url.start_with?("//")
|
||||||
raise Discourse::InvalidParameters.new(:url) unless url =~ /^https?:\/\//
|
raise Discourse::InvalidParameters.new(:url) unless url =~ /^https?:\/\//
|
||||||
|
|
||||||
dest = FinalDestination.new(
|
tmp = nil
|
||||||
|
|
||||||
|
fd = FinalDestination.new(
|
||||||
url,
|
url,
|
||||||
max_redirects: follow_redirect ? 5 : 1,
|
max_redirects: follow_redirect ? 5 : 1,
|
||||||
skip_rate_limit: skip_rate_limit,
|
skip_rate_limit: skip_rate_limit,
|
||||||
verbose: verbose
|
verbose: verbose
|
||||||
)
|
)
|
||||||
uri = dest.resolve
|
|
||||||
|
|
||||||
if !uri && dest.status_code.to_i >= 400
|
fd.get do |response, chunk, uri|
|
||||||
# attempt error API compatability
|
if tmp.nil?
|
||||||
|
# error handling
|
||||||
|
if uri.blank?
|
||||||
|
if response.code.to_i >= 400
|
||||||
|
# attempt error API compatibility
|
||||||
io = FakeIO.new
|
io = FakeIO.new
|
||||||
io.status = [dest.status_code.to_s, ""]
|
io.status = [response.code, ""]
|
||||||
|
raise OpenURI::HTTPError.new("#{response.code} Error", io)
|
||||||
# TODO perhaps translate and add Discourse::DownloadError
|
else
|
||||||
raise OpenURI::HTTPError.new("#{dest.status_code} Error", io)
|
|
||||||
end
|
|
||||||
|
|
||||||
unless uri
|
|
||||||
log(:error, "FinalDestination did not work for: #{url}") if verbose
|
log(:error, "FinalDestination did not work for: #{url}") if verbose
|
||||||
return
|
throw :done
|
||||||
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
downloaded = uri.open("rb", read_timeout: read_timeout)
|
# first run
|
||||||
|
tmp_file_ext = File.extname(uri.path)
|
||||||
|
|
||||||
extension = File.extname(uri.path)
|
if tmp_file_ext.blank? && response.content_type.present?
|
||||||
|
ext = MiniMime.lookup_by_content_type(response.content_type)&.extension
|
||||||
if extension.blank? && downloaded.content_type.present?
|
|
||||||
ext = MiniMime.lookup_by_content_type(downloaded.content_type)&.extension
|
|
||||||
ext = "jpg" if ext == "jpe"
|
ext = "jpg" if ext == "jpe"
|
||||||
extension = "." + ext if ext.present?
|
tmp_file_ext = "." + ext if ext.present?
|
||||||
end
|
end
|
||||||
|
|
||||||
tmp = Tempfile.new([tmp_file_name, extension])
|
tmp = Tempfile.new([tmp_file_name, tmp_file_ext])
|
||||||
|
tmp.binmode
|
||||||
File.open(tmp.path, "wb") do |f|
|
|
||||||
while f.size <= max_file_size && data = downloaded.read(512.kilobytes)
|
|
||||||
f.write(data)
|
|
||||||
end
|
|
||||||
end
|
end
|
||||||
|
|
||||||
|
tmp.write(chunk)
|
||||||
|
|
||||||
|
throw :done if tmp.size > max_file_size
|
||||||
|
end
|
||||||
|
|
||||||
|
tmp&.rewind
|
||||||
tmp
|
tmp
|
||||||
ensure
|
|
||||||
downloaded&.close
|
|
||||||
end
|
end
|
||||||
|
|
||||||
def self.optimize_image!(filename)
|
def self.optimize_image!(filename)
|
||||||
|
|
|
@ -96,9 +96,7 @@ class FinalDestination
|
||||||
uri = URI(uri.to_s)
|
uri = URI(uri.to_s)
|
||||||
end
|
end
|
||||||
|
|
||||||
unless validate_uri
|
return nil unless validate_uri
|
||||||
return nil
|
|
||||||
end
|
|
||||||
|
|
||||||
result, (location, cookie) = safe_get(uri, &blk)
|
result, (location, cookie) = safe_get(uri, &blk)
|
||||||
|
|
||||||
|
@ -120,9 +118,7 @@ class FinalDestination
|
||||||
return nil if !uri
|
return nil if !uri
|
||||||
|
|
||||||
extra = nil
|
extra = nil
|
||||||
if cookie
|
extra = { 'Cookie' => cookie } if cookie
|
||||||
extra = { 'Cookie' => cookie }
|
|
||||||
end
|
|
||||||
|
|
||||||
get(uri, redirects - 1, extra_headers: extra, &blk)
|
get(uri, redirects - 1, extra_headers: extra, &blk)
|
||||||
elsif result == :ok
|
elsif result == :ok
|
||||||
|
@ -251,7 +247,6 @@ class FinalDestination
|
||||||
end
|
end
|
||||||
|
|
||||||
def is_dest_valid?
|
def is_dest_valid?
|
||||||
|
|
||||||
return false unless @uri && @uri.host
|
return false unless @uri && @uri.host
|
||||||
|
|
||||||
# Whitelisted hosts
|
# Whitelisted hosts
|
||||||
|
@ -260,9 +255,7 @@ class FinalDestination
|
||||||
hostname_matches?(Discourse.base_url_no_prefix)
|
hostname_matches?(Discourse.base_url_no_prefix)
|
||||||
|
|
||||||
if SiteSetting.whitelist_internal_hosts.present?
|
if SiteSetting.whitelist_internal_hosts.present?
|
||||||
SiteSetting.whitelist_internal_hosts.split('|').each do |h|
|
return true if SiteSetting.whitelist_internal_hosts.split("|").any? { |h| h.downcase == @uri.hostname.downcase }
|
||||||
return true if @uri.hostname.downcase == h.downcase
|
|
||||||
end
|
|
||||||
end
|
end
|
||||||
|
|
||||||
address_s = @opts[:lookup_ip].call(@uri.hostname)
|
address_s = @opts[:lookup_ip].call(@uri.hostname)
|
||||||
|
@ -331,12 +324,10 @@ class FinalDestination
|
||||||
protected
|
protected
|
||||||
|
|
||||||
def safe_get(uri)
|
def safe_get(uri)
|
||||||
|
|
||||||
result = nil
|
result = nil
|
||||||
unsafe_close = false
|
unsafe_close = false
|
||||||
|
|
||||||
safe_session(uri) do |http|
|
safe_session(uri) do |http|
|
||||||
|
|
||||||
headers = request_headers.merge(
|
headers = request_headers.merge(
|
||||||
'Accept-Encoding' => 'gzip',
|
'Accept-Encoding' => 'gzip',
|
||||||
'Host' => uri.host
|
'Host' => uri.host
|
||||||
|
@ -350,9 +341,8 @@ class FinalDestination
|
||||||
end
|
end
|
||||||
|
|
||||||
if Net::HTTPSuccess === resp
|
if Net::HTTPSuccess === resp
|
||||||
|
|
||||||
resp.decode_content = true
|
resp.decode_content = true
|
||||||
resp.read_body { |chunk|
|
resp.read_body do |chunk|
|
||||||
read_next = true
|
read_next = true
|
||||||
|
|
||||||
catch(:done) do
|
catch(:done) do
|
||||||
|
@ -370,19 +360,19 @@ class FinalDestination
|
||||||
http.finish
|
http.finish
|
||||||
raise StandardError
|
raise StandardError
|
||||||
end
|
end
|
||||||
}
|
end
|
||||||
result = :ok
|
result = :ok
|
||||||
|
else
|
||||||
|
catch(:done) do
|
||||||
|
yield resp, nil, nil
|
||||||
|
end
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
result
|
result
|
||||||
rescue StandardError
|
rescue StandardError
|
||||||
if unsafe_close
|
unsafe_close ? :ok : raise
|
||||||
:ok
|
|
||||||
else
|
|
||||||
raise
|
|
||||||
end
|
|
||||||
end
|
end
|
||||||
|
|
||||||
def safe_session(uri)
|
def safe_session(uri)
|
||||||
|
|
|
@ -507,8 +507,8 @@ describe CookedPostProcessor do
|
||||||
</html>
|
</html>
|
||||||
HTML
|
HTML
|
||||||
|
|
||||||
stub_request(:head, url).to_return(status: 200)
|
stub_request(:head, url)
|
||||||
stub_request(:get , url).to_return(status: 200, body: body)
|
stub_request(:get , url).to_return(body: body)
|
||||||
FinalDestination.stubs(:lookup_ip).returns('1.2.3.4')
|
FinalDestination.stubs(:lookup_ip).returns('1.2.3.4')
|
||||||
|
|
||||||
# not an ideal stub but shipping the whole image to fast image can add
|
# not an ideal stub but shipping the whole image to fast image can add
|
||||||
|
|
|
@ -15,7 +15,6 @@ describe FileHelper do
|
||||||
|
|
||||||
it "correctly raises an OpenURI HTTP error if it gets a 404 even with redirect" do
|
it "correctly raises an OpenURI HTTP error if it gets a 404 even with redirect" do
|
||||||
url = "http://fourohfour.com/404"
|
url = "http://fourohfour.com/404"
|
||||||
stub_request(:head, url).to_return(status: 404, body: "404")
|
|
||||||
stub_request(:get, url).to_return(status: 404, body: "404")
|
stub_request(:get, url).to_return(status: 404, body: "404")
|
||||||
|
|
||||||
expect do
|
expect do
|
||||||
|
@ -36,7 +35,6 @@ describe FileHelper do
|
||||||
it "correctly raises an OpenURI HTTP error if it gets a 404" do
|
it "correctly raises an OpenURI HTTP error if it gets a 404" do
|
||||||
url = "http://fourohfour.com/404"
|
url = "http://fourohfour.com/404"
|
||||||
|
|
||||||
stub_request(:head, url).to_return(status: 404, body: "404")
|
|
||||||
stub_request(:get, url).to_return(status: 404, body: "404")
|
stub_request(:get, url).to_return(status: 404, body: "404")
|
||||||
|
|
||||||
expect do
|
expect do
|
||||||
|
|
|
@ -176,7 +176,7 @@ describe FinalDestination do
|
||||||
'Set-Cookie' => 'evil=trout'
|
'Set-Cookie' => 'evil=trout'
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
stub_request(:head, 'https://discourse.org').to_return(status: 200)
|
stub_request(:head, 'https://discourse.org')
|
||||||
end
|
end
|
||||||
|
|
||||||
context "when the status code is 405" do
|
context "when the status code is 405" do
|
||||||
|
@ -218,7 +218,6 @@ describe FinalDestination do
|
||||||
|
|
||||||
stub_request(:head, "https://eviltrout.com")
|
stub_request(:head, "https://eviltrout.com")
|
||||||
.with(headers: { "Cookie" => "bar=1; baz=2; foo=219ffwef9w0f" })
|
.with(headers: { "Cookie" => "bar=1; baz=2; foo=219ffwef9w0f" })
|
||||||
.to_return(status: 200, body: "")
|
|
||||||
|
|
||||||
final = FinalDestination.new("https://eviltrout.com", opts)
|
final = FinalDestination.new("https://eviltrout.com", opts)
|
||||||
expect(final.resolve.to_s).to eq("https://eviltrout.com")
|
expect(final.resolve.to_s).to eq("https://eviltrout.com")
|
||||||
|
@ -229,14 +228,13 @@ describe FinalDestination do
|
||||||
|
|
||||||
it "should use the correct format for cookies when there is only one cookie" do
|
it "should use the correct format for cookies when there is only one cookie" do
|
||||||
stub_request(:head, "https://eviltrout.com")
|
stub_request(:head, "https://eviltrout.com")
|
||||||
.to_return(status: 302, body: "" , headers: {
|
.to_return(status: 302, headers: {
|
||||||
"Location" => "https://eviltrout.com",
|
"Location" => "https://eviltrout.com",
|
||||||
"Set-Cookie" => "foo=219ffwef9w0f; expires=Mon, 19-Feb-2018 10:44:24 GMT; path=/; domain=eviltrout.com"
|
"Set-Cookie" => "foo=219ffwef9w0f; expires=Mon, 19-Feb-2018 10:44:24 GMT; path=/; domain=eviltrout.com"
|
||||||
})
|
})
|
||||||
|
|
||||||
stub_request(:head, "https://eviltrout.com")
|
stub_request(:head, "https://eviltrout.com")
|
||||||
.with(headers: { "Cookie" => "foo=219ffwef9w0f" })
|
.with(headers: { "Cookie" => "foo=219ffwef9w0f" })
|
||||||
.to_return(status: 200, body: "")
|
|
||||||
|
|
||||||
final = FinalDestination.new("https://eviltrout.com", opts)
|
final = FinalDestination.new("https://eviltrout.com", opts)
|
||||||
expect(final.resolve.to_s).to eq("https://eviltrout.com")
|
expect(final.resolve.to_s).to eq("https://eviltrout.com")
|
||||||
|
@ -246,7 +244,7 @@ describe FinalDestination do
|
||||||
|
|
||||||
it "should use the correct format for cookies when there are multiple cookies" do
|
it "should use the correct format for cookies when there are multiple cookies" do
|
||||||
stub_request(:head, "https://eviltrout.com")
|
stub_request(:head, "https://eviltrout.com")
|
||||||
.to_return(status: 302, body: "" , headers: {
|
.to_return(status: 302, headers: {
|
||||||
"Location" => "https://eviltrout.com",
|
"Location" => "https://eviltrout.com",
|
||||||
"Set-Cookie" => ["foo=219ffwef9w0f; expires=Mon, 19-Feb-2018 10:44:24 GMT; path=/; domain=eviltrout.com",
|
"Set-Cookie" => ["foo=219ffwef9w0f; expires=Mon, 19-Feb-2018 10:44:24 GMT; path=/; domain=eviltrout.com",
|
||||||
"bar=1",
|
"bar=1",
|
||||||
|
@ -255,7 +253,6 @@ describe FinalDestination do
|
||||||
|
|
||||||
stub_request(:head, "https://eviltrout.com")
|
stub_request(:head, "https://eviltrout.com")
|
||||||
.with(headers: { "Cookie" => "bar=1; baz=2; foo=219ffwef9w0f" })
|
.with(headers: { "Cookie" => "bar=1; baz=2; foo=219ffwef9w0f" })
|
||||||
.to_return(status: 200, body: "")
|
|
||||||
|
|
||||||
final = FinalDestination.new("https://eviltrout.com", opts)
|
final = FinalDestination.new("https://eviltrout.com", opts)
|
||||||
expect(final.resolve.to_s).to eq("https://eviltrout.com")
|
expect(final.resolve.to_s).to eq("https://eviltrout.com")
|
||||||
|
@ -267,7 +264,6 @@ describe FinalDestination do
|
||||||
describe '.get' do
|
describe '.get' do
|
||||||
|
|
||||||
it "can correctly stream with a redirect" do
|
it "can correctly stream with a redirect" do
|
||||||
|
|
||||||
FinalDestination.clear_https_cache!("wikipedia.com")
|
FinalDestination.clear_https_cache!("wikipedia.com")
|
||||||
|
|
||||||
stub_request(:get, "http://wikipedia.com/").
|
stub_request(:get, "http://wikipedia.com/").
|
||||||
|
@ -399,13 +395,12 @@ describe FinalDestination do
|
||||||
|
|
||||||
stub_request(:head, "http://wikipedia.com/image.png")
|
stub_request(:head, "http://wikipedia.com/image.png")
|
||||||
.to_return(status: 302, body: "", headers: { location: 'https://wikipedia.com/image.png' })
|
.to_return(status: 302, body: "", headers: { location: 'https://wikipedia.com/image.png' })
|
||||||
|
|
||||||
stub_request(:head, "https://wikipedia.com/image.png")
|
stub_request(:head, "https://wikipedia.com/image.png")
|
||||||
.to_return(status: 200, body: "", headers: [])
|
|
||||||
|
|
||||||
fd('http://wikipedia.com/image.png').resolve
|
fd('http://wikipedia.com/image.png').resolve
|
||||||
|
|
||||||
stub_request(:head, "https://wikipedia.com/image2.png")
|
stub_request(:head, "https://wikipedia.com/image2.png")
|
||||||
.to_return(status: 200, body: "", headers: [])
|
|
||||||
|
|
||||||
fd('http://wikipedia.com/image2.png').resolve
|
fd('http://wikipedia.com/image2.png').resolve
|
||||||
end
|
end
|
||||||
|
|
|
@ -71,11 +71,8 @@ describe InlineOneboxer do
|
||||||
it "will crawl anything if allowed to" do
|
it "will crawl anything if allowed to" do
|
||||||
SiteSetting.enable_inline_onebox_on_all_domains = true
|
SiteSetting.enable_inline_onebox_on_all_domains = true
|
||||||
|
|
||||||
# Final destination does a HEAD and a GET
|
|
||||||
stub_request(:head, "https://eviltrout.com/some-path").to_return(status: 200)
|
|
||||||
|
|
||||||
stub_request(:get, "https://eviltrout.com/some-path").
|
stub_request(:get, "https://eviltrout.com/some-path").
|
||||||
to_return(status: 200, body: "<html><head><title>a blog</title></head></html>", headers: {})
|
to_return(status: 200, body: "<html><head><title>a blog</title></head></html>")
|
||||||
|
|
||||||
onebox = InlineOneboxer.lookup(
|
onebox = InlineOneboxer.lookup(
|
||||||
"https://eviltrout.com/some-path",
|
"https://eviltrout.com/some-path",
|
||||||
|
@ -90,11 +87,8 @@ describe InlineOneboxer do
|
||||||
it "will not return a onebox if it does not meet minimal length" do
|
it "will not return a onebox if it does not meet minimal length" do
|
||||||
SiteSetting.enable_inline_onebox_on_all_domains = true
|
SiteSetting.enable_inline_onebox_on_all_domains = true
|
||||||
|
|
||||||
# Final destination does a HEAD and a GET
|
|
||||||
stub_request(:head, "https://eviltrout.com/some-path").to_return(status: 200)
|
|
||||||
|
|
||||||
stub_request(:get, "https://eviltrout.com/some-path").
|
stub_request(:get, "https://eviltrout.com/some-path").
|
||||||
to_return(status: 200, body: "<html><head><title>a</title></head></html>", headers: {})
|
to_return(status: 200, body: "<html><head><title>a</title></head></html>")
|
||||||
|
|
||||||
onebox = InlineOneboxer.lookup(
|
onebox = InlineOneboxer.lookup(
|
||||||
"https://eviltrout.com/some-path",
|
"https://eviltrout.com/some-path",
|
||||||
|
|
|
@ -4,8 +4,8 @@ require_dependency 'oneboxer'
|
||||||
describe Oneboxer do
|
describe Oneboxer do
|
||||||
|
|
||||||
it "returns blank string for an invalid onebox" do
|
it "returns blank string for an invalid onebox" do
|
||||||
|
stub_request(:head, "http://boom.com")
|
||||||
stub_request(:get, "http://boom.com").to_return(body: "")
|
stub_request(:get, "http://boom.com").to_return(body: "")
|
||||||
stub_request(:head, "http://boom.com").to_return(body: "")
|
|
||||||
|
|
||||||
expect(Oneboxer.preview("http://boom.com")).to eq("")
|
expect(Oneboxer.preview("http://boom.com")).to eq("")
|
||||||
expect(Oneboxer.onebox("http://boom.com")).to eq("")
|
expect(Oneboxer.onebox("http://boom.com")).to eq("")
|
||||||
|
|
|
@ -952,10 +952,8 @@ HTML
|
||||||
SiteSetting.enable_inline_onebox_on_all_domains = true
|
SiteSetting.enable_inline_onebox_on_all_domains = true
|
||||||
InlineOneboxer.purge("http://cnn.com/a")
|
InlineOneboxer.purge("http://cnn.com/a")
|
||||||
|
|
||||||
stub_request(:head, "http://cnn.com/a").to_return(status: 200)
|
|
||||||
|
|
||||||
stub_request(:get, "http://cnn.com/a").
|
stub_request(:get, "http://cnn.com/a").
|
||||||
to_return(status: 200, body: "<html><head><title>news</title></head></html>", headers: {})
|
to_return(status: 200, body: "<html><head><title>news</title></head></html>")
|
||||||
|
|
||||||
expect(PrettyText.cook("- http://cnn.com/a\n- a http://cnn.com/a").split("news").length).to eq(3)
|
expect(PrettyText.cook("- http://cnn.com/a\n- a http://cnn.com/a").split("news").length).to eq(3)
|
||||||
expect(PrettyText.cook("- http://cnn.com/a\n - a http://cnn.com/a").split("news").length).to eq(3)
|
expect(PrettyText.cook("- http://cnn.com/a\n - a http://cnn.com/a").split("news").length).to eq(3)
|
||||||
|
@ -965,10 +963,8 @@ HTML
|
||||||
SiteSetting.enable_inline_onebox_on_all_domains = true
|
SiteSetting.enable_inline_onebox_on_all_domains = true
|
||||||
InlineOneboxer.purge("http://cnn.com?a")
|
InlineOneboxer.purge("http://cnn.com?a")
|
||||||
|
|
||||||
stub_request(:head, "http://cnn.com?a").to_return(status: 200)
|
|
||||||
|
|
||||||
stub_request(:get, "http://cnn.com?a").
|
stub_request(:get, "http://cnn.com?a").
|
||||||
to_return(status: 200, body: "<html><head><title>news</title></head></html>", headers: {})
|
to_return(status: 200, body: "<html><head><title>news</title></head></html>")
|
||||||
|
|
||||||
expect(PrettyText.cook("- http://cnn.com?a\n- a http://cnn.com?a").split("news").length).to eq(3)
|
expect(PrettyText.cook("- http://cnn.com?a\n- a http://cnn.com?a").split("news").length).to eq(3)
|
||||||
expect(PrettyText.cook("- http://cnn.com?a\n - a http://cnn.com?a").split("news").length).to eq(3)
|
expect(PrettyText.cook("- http://cnn.com?a\n - a http://cnn.com?a").split("news").length).to eq(3)
|
||||||
|
@ -981,9 +977,8 @@ HTML
|
||||||
SiteSetting.enable_inline_onebox_on_all_domains = true
|
SiteSetting.enable_inline_onebox_on_all_domains = true
|
||||||
InlineOneboxer.purge("http://cnn.com/")
|
InlineOneboxer.purge("http://cnn.com/")
|
||||||
|
|
||||||
stub_request(:head, "http://cnn.com/").to_return(status: 200)
|
|
||||||
stub_request(:get, "http://cnn.com/").
|
stub_request(:get, "http://cnn.com/").
|
||||||
to_return(status: 200, body: "<html><head><title>news</title></head></html>", headers: {})
|
to_return(status: 200, body: "<html><head><title>news</title></head></html>")
|
||||||
|
|
||||||
expect(PrettyText.cook("- http://cnn.com/\n- a http://cnn.com/").split("news").length).to eq(1)
|
expect(PrettyText.cook("- http://cnn.com/\n- a http://cnn.com/").split("news").length).to eq(1)
|
||||||
expect(PrettyText.cook("- cnn.com\n - a http://cnn.com/").split("news").length).to eq(1)
|
expect(PrettyText.cook("- cnn.com\n - a http://cnn.com/").split("news").length).to eq(1)
|
||||||
|
|
|
@ -35,11 +35,8 @@ describe OneboxController do
|
||||||
|
|
||||||
url = "http://noodle.com/"
|
url = "http://noodle.com/"
|
||||||
|
|
||||||
stub_request(:head, url).
|
stub_request(:head, url)
|
||||||
to_return(status: 200, body: "", headers: {}).then.to_raise
|
stub_request(:get, url).to_return(body: onebox_html).then.to_raise
|
||||||
|
|
||||||
stub_request(:get, url)
|
|
||||||
.to_return(status: 200, headers: {}, body: onebox_html).then.to_raise
|
|
||||||
|
|
||||||
get :show, params: { url: url, refresh: "true" }, format: :json
|
get :show, params: { url: url, refresh: "true" }, format: :json
|
||||||
|
|
||||||
|
|
|
@ -59,21 +59,20 @@ describe UploadsController do
|
||||||
expect(id).to be
|
expect(id).to be
|
||||||
end
|
end
|
||||||
|
|
||||||
it 'is successful with synchronous api' do
|
it 'is successful with api' do
|
||||||
SiteSetting.authorized_extensions = "*"
|
SiteSetting.authorized_extensions = "*"
|
||||||
controller.stubs(:is_api?).returns(true)
|
controller.stubs(:is_api?).returns(true)
|
||||||
|
|
||||||
|
FinalDestination.stubs(:lookup_ip).returns("1.2.3.4")
|
||||||
|
|
||||||
Jobs.expects(:enqueue).with(:create_avatar_thumbnails, anything)
|
Jobs.expects(:enqueue).with(:create_avatar_thumbnails, anything)
|
||||||
|
|
||||||
stub_request(:head, 'http://example.com/image.png')
|
url = "http://example.com/image.png"
|
||||||
stub_request(:get, "http://example.com/image.png").to_return(body: File.read('spec/fixtures/images/logo.png'))
|
png = File.read(Rails.root + "spec/fixtures/images/logo.png")
|
||||||
|
|
||||||
post :create, params: {
|
stub_request(:get, url).to_return(status: 200, body: png)
|
||||||
url: 'http://example.com/image.png',
|
|
||||||
type: "avatar",
|
post :create, params: { url: url, type: "avatar", format: :json }
|
||||||
synchronous: true,
|
|
||||||
format: :json
|
|
||||||
}
|
|
||||||
|
|
||||||
json = ::JSON.parse(response.body)
|
json = ::JSON.parse(response.body)
|
||||||
|
|
||||||
|
|
|
@ -31,7 +31,6 @@ describe UserAvatarsController do
|
||||||
SiteSetting.s3_upload_bucket = "test"
|
SiteSetting.s3_upload_bucket = "test"
|
||||||
SiteSetting.s3_cdn_url = "http://cdn.com"
|
SiteSetting.s3_cdn_url = "http://cdn.com"
|
||||||
|
|
||||||
stub_request(:head, "http://cdn.com/something/else")
|
|
||||||
stub_request(:get, "http://cdn.com/something/else").to_return(body: 'image')
|
stub_request(:get, "http://cdn.com/something/else").to_return(body: 'image')
|
||||||
|
|
||||||
GlobalSetting.expects(:cdn_url).returns("http://awesome.com/boom")
|
GlobalSetting.expects(:cdn_url).returns("http://awesome.com/boom")
|
||||||
|
|
|
@ -99,9 +99,8 @@ describe Jobs::PollFeed do
|
||||||
SiteSetting.feed_polling_url = 'https://blog.discourse.org/feed/'
|
SiteSetting.feed_polling_url = 'https://blog.discourse.org/feed/'
|
||||||
SiteSetting.embed_by_username = embed_by_username
|
SiteSetting.embed_by_username = embed_by_username
|
||||||
|
|
||||||
stub_request(:head, SiteSetting.feed_polling_url).to_return(status: 200)
|
stub_request(:head, SiteSetting.feed_polling_url)
|
||||||
stub_request(:get, SiteSetting.feed_polling_url).to_return(
|
stub_request(:get, SiteSetting.feed_polling_url).to_return(
|
||||||
status: 200,
|
|
||||||
body: file_from_fixtures('feed.rss', 'feed').read,
|
body: file_from_fixtures('feed.rss', 'feed').read,
|
||||||
headers: { "Content-Type" => "application/rss+xml" }
|
headers: { "Content-Type" => "application/rss+xml" }
|
||||||
)
|
)
|
||||||
|
@ -116,9 +115,8 @@ describe Jobs::PollFeed do
|
||||||
SiteSetting.feed_polling_url = 'https://blog.discourse.org/feed/atom/'
|
SiteSetting.feed_polling_url = 'https://blog.discourse.org/feed/atom/'
|
||||||
SiteSetting.embed_by_username = embed_by_username
|
SiteSetting.embed_by_username = embed_by_username
|
||||||
|
|
||||||
stub_request(:head, SiteSetting.feed_polling_url).to_return(status: 200)
|
stub_request(:head, SiteSetting.feed_polling_url)
|
||||||
stub_request(:get, SiteSetting.feed_polling_url).to_return(
|
stub_request(:get, SiteSetting.feed_polling_url).to_return(
|
||||||
status: 200,
|
|
||||||
body: file_from_fixtures('feed.atom', 'feed').read,
|
body: file_from_fixtures('feed.atom', 'feed').read,
|
||||||
headers: { "Content-Type" => "application/atom+xml" }
|
headers: { "Content-Type" => "application/atom+xml" }
|
||||||
)
|
)
|
||||||
|
|
|
@ -11,10 +11,8 @@ describe Jobs::PullHotlinkedImages do
|
||||||
|
|
||||||
before do
|
before do
|
||||||
stub_request(:get, image_url).to_return(body: png, headers: { "Content-Type" => "image/png" })
|
stub_request(:get, image_url).to_return(body: png, headers: { "Content-Type" => "image/png" })
|
||||||
stub_request(:head, image_url)
|
stub_request(:get, broken_image_url).to_return(status: 404)
|
||||||
stub_request(:head, broken_image_url).to_return(status: 404)
|
|
||||||
stub_request(:get, large_image_url).to_return(body: large_png, headers: { "Content-Type" => "image/png" })
|
stub_request(:get, large_image_url).to_return(body: large_png, headers: { "Content-Type" => "image/png" })
|
||||||
stub_request(:head, large_image_url)
|
|
||||||
SiteSetting.crawl_images = true
|
SiteSetting.crawl_images = true
|
||||||
SiteSetting.download_remote_images_to_local = true
|
SiteSetting.download_remote_images_to_local = true
|
||||||
SiteSetting.max_image_size_kb = 2
|
SiteSetting.max_image_size_kb = 2
|
||||||
|
@ -59,7 +57,6 @@ describe Jobs::PullHotlinkedImages do
|
||||||
it 'replaces images without extension' do
|
it 'replaces images without extension' do
|
||||||
url = image_url.sub(/\.[a-zA-Z0-9]+$/, '')
|
url = image_url.sub(/\.[a-zA-Z0-9]+$/, '')
|
||||||
stub_request(:get, url).to_return(body: png, headers: { "Content-Type" => "image/png" })
|
stub_request(:get, url).to_return(body: png, headers: { "Content-Type" => "image/png" })
|
||||||
stub_request(:head, url)
|
|
||||||
post = Fabricate(:post, raw: "<img src='#{url}'>")
|
post = Fabricate(:post, raw: "<img src='#{url}'>")
|
||||||
|
|
||||||
Jobs::PullHotlinkedImages.new.execute(post_id: post.id)
|
Jobs::PullHotlinkedImages.new.execute(post_id: post.id)
|
||||||
|
@ -75,8 +72,8 @@ describe Jobs::PullHotlinkedImages do
|
||||||
|
|
||||||
before do
|
before do
|
||||||
SiteSetting.queue_jobs = true
|
SiteSetting.queue_jobs = true
|
||||||
stub_request(:get, url).to_return(body: '')
|
|
||||||
stub_request(:head, url)
|
stub_request(:head, url)
|
||||||
|
stub_request(:get, url).to_return(body: '')
|
||||||
stub_request(:get, api_url).to_return(body: "{
|
stub_request(:get, api_url).to_return(body: "{
|
||||||
\"query\": {
|
\"query\": {
|
||||||
\"pages\": {
|
\"pages\": {
|
||||||
|
@ -91,7 +88,6 @@ describe Jobs::PullHotlinkedImages do
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}")
|
}")
|
||||||
stub_request(:head, api_url)
|
|
||||||
end
|
end
|
||||||
|
|
||||||
it 'replaces image src' do
|
it 'replaces image src' do
|
||||||
|
|
|
@ -9,7 +9,6 @@ describe Jobs::UpdateGravatar do
|
||||||
|
|
||||||
png = Base64.decode64("R0lGODlhAQABALMAAAAAAIAAAACAAICAAAAAgIAAgACAgMDAwICAgP8AAAD/AP//AAAA//8A/wD//wBiZCH5BAEAAA8ALAAAAAABAAEAAAQC8EUAOw==")
|
png = Base64.decode64("R0lGODlhAQABALMAAAAAAIAAAACAAICAAAAAgIAAgACAgMDAwICAgP8AAAD/AP//AAAA//8A/wD//wBiZCH5BAEAAA8ALAAAAAABAAEAAAQC8EUAOw==")
|
||||||
url = "https://www.gravatar.com/avatar/d10ca8d11301c2f4993ac2279ce4b930.png?s=360&d=404"
|
url = "https://www.gravatar.com/avatar/d10ca8d11301c2f4993ac2279ce4b930.png?s=360&d=404"
|
||||||
stub_request(:head, url).to_return(status: 200)
|
|
||||||
stub_request(:get, url).to_return(body: png)
|
stub_request(:get, url).to_return(body: png)
|
||||||
|
|
||||||
SiteSetting.automatically_download_gravatars = true
|
SiteSetting.automatically_download_gravatars = true
|
||||||
|
|
|
@ -47,7 +47,7 @@ describe UserAvatar do
|
||||||
describe 'when avatar url returns an invalid status code' do
|
describe 'when avatar url returns an invalid status code' do
|
||||||
it 'should not do anything' do
|
it 'should not do anything' do
|
||||||
url = "http://thisfakesomething.something.com/"
|
url = "http://thisfakesomething.something.com/"
|
||||||
stub_request(:head, url).to_return(status: 404)
|
|
||||||
UserAvatar.import_url_for_user(url, user)
|
UserAvatar.import_url_for_user(url, user)
|
||||||
|
|
||||||
user.reload
|
user.reload
|
||||||
|
|
|
@ -3,20 +3,14 @@ require 'rails_helper'
|
||||||
describe StaticController do
|
describe StaticController do
|
||||||
|
|
||||||
context '#favicon' do
|
context '#favicon' do
|
||||||
before do
|
|
||||||
# this is a mess in test, will fix in a future commit
|
|
||||||
FinalDestination.stubs(:lookup_ip).returns('1.2.3.4')
|
|
||||||
end
|
|
||||||
|
|
||||||
let(:png) { Base64.decode64("R0lGODlhAQABALMAAAAAAIAAAACAAICAAAAAgIAAgACAgMDAwICAgP8AAAD/AP//AAAA//8A/wD//wBiZCH5BAEAAA8ALAAAAAABAAEAAAQC8EUAOw==") }
|
let(:png) { Base64.decode64("R0lGODlhAQABALMAAAAAAIAAAACAAICAAAAAgIAAgACAgMDAwICAgP8AAAD/AP//AAAA//8A/wD//wBiZCH5BAEAAA8ALAAAAAABAAEAAAQC8EUAOw==") }
|
||||||
|
|
||||||
|
before { FinalDestination.stubs(:lookup_ip).returns("1.2.3.4") }
|
||||||
|
|
||||||
it 'returns the default favicon for a missing download' do
|
it 'returns the default favicon for a missing download' do
|
||||||
|
url = "https://fav.icon/#{SecureRandom.hex}.png"
|
||||||
|
|
||||||
url = "https://somewhere1.over.rainbow/#{SecureRandom.hex}.png"
|
stub_request(:get, url).to_return(status: 404)
|
||||||
|
|
||||||
stub_request(:head, url).
|
|
||||||
with(headers: { 'Host' => 'somewhere1.over.rainbow' }).
|
|
||||||
to_return(status: 404, body: "", headers: {})
|
|
||||||
|
|
||||||
SiteSetting.favicon_url = url
|
SiteSetting.favicon_url = url
|
||||||
|
|
||||||
|
@ -30,14 +24,9 @@ describe StaticController do
|
||||||
end
|
end
|
||||||
|
|
||||||
it 'can proxy a favicon correctly' do
|
it 'can proxy a favicon correctly' do
|
||||||
url = "https://somewhere.over.rainbow/#{SecureRandom.hex}.png"
|
url = "https://fav.icon/#{SecureRandom.hex}.png"
|
||||||
|
|
||||||
stub_request(:head, url).
|
stub_request(:get, url).to_return(status: 200, body: png)
|
||||||
with(headers: { 'Host' => 'somewhere.over.rainbow' }).
|
|
||||||
to_return(status: 200, body: "", headers: {})
|
|
||||||
|
|
||||||
stub_request(:get, url).
|
|
||||||
to_return(status: 200, body: png, headers: {})
|
|
||||||
|
|
||||||
SiteSetting.favicon_url = url
|
SiteSetting.favicon_url = url
|
||||||
|
|
||||||
|
@ -51,7 +40,6 @@ describe StaticController do
|
||||||
|
|
||||||
context '#brotli_asset' do
|
context '#brotli_asset' do
|
||||||
it 'returns a non brotli encoded 404 if asset is missing' do
|
it 'returns a non brotli encoded 404 if asset is missing' do
|
||||||
|
|
||||||
get "/brotli_asset/missing.js"
|
get "/brotli_asset/missing.js"
|
||||||
|
|
||||||
expect(response.status).to eq(404)
|
expect(response.status).to eq(404)
|
||||||
|
|
Loading…
Reference in New Issue