2016-02-24 12:57:29 +01:00
|
|
|
class ProcessFeedService < BaseService
|
2016-10-10 18:05:52 +02:00
|
|
|
ACTIVITY_NS = 'http://activitystrea.ms/spec/1.0/'.freeze
|
|
|
|
THREAD_NS = 'http://purl.org/syndication/thread/1.0'.freeze
|
|
|
|
|
2016-02-20 22:53:20 +01:00
|
|
|
def call(body, account)
|
|
|
|
xml = Nokogiri::XML(body)
|
2016-11-08 01:32:34 +01:00
|
|
|
|
|
|
|
update_author(xml, account)
|
|
|
|
process_entries(xml, account)
|
2016-03-25 02:13:30 +01:00
|
|
|
end
|
2016-02-20 22:53:20 +01:00
|
|
|
|
2016-03-25 02:13:30 +01:00
|
|
|
private
|
2016-02-28 14:26:26 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
def update_author(xml, account)
|
|
|
|
return if xml.at_xpath('/xmlns:feed').nil?
|
|
|
|
UpdateRemoteProfileService.new.call(xml.at_xpath('/xmlns:feed/xmlns:author'), account)
|
|
|
|
end
|
2016-02-24 01:28:53 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
def process_entries(xml, account)
|
|
|
|
xml.xpath('//xmlns:entry').reverse_each.map { |entry| ProcessEntry.new.call(entry, account) }.compact
|
|
|
|
end
|
2016-03-16 10:46:15 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
class ProcessEntry
|
|
|
|
def call(xml, account)
|
|
|
|
@account = account
|
|
|
|
@xml = xml
|
2016-02-24 01:28:53 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
return if skip_unsupported_type?
|
2016-02-24 01:28:53 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
case verb
|
|
|
|
when :post, :share
|
|
|
|
return create_status
|
|
|
|
when :delete
|
|
|
|
return delete_status
|
2016-02-24 01:28:53 +01:00
|
|
|
end
|
2016-03-25 02:13:30 +01:00
|
|
|
end
|
2016-02-25 00:17:01 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
private
|
2016-09-09 20:04:34 +02:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
def create_status
|
|
|
|
Rails.logger.debug "Creating remote status #{id}"
|
|
|
|
status = status_from_xml(@xml)
|
2016-11-05 15:20:05 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
if verb == :share
|
2016-11-08 01:48:17 +01:00
|
|
|
original_status = status_from_xml(@xml.at_xpath('.//activity:object', activity: ACTIVITY_NS))
|
2016-11-08 01:32:34 +01:00
|
|
|
status.reblog = original_status
|
|
|
|
end
|
2016-09-09 20:04:34 +02:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
status.save!
|
|
|
|
Rails.logger.debug "Queuing remote status #{status.id} (#{id}) for distribution"
|
2016-03-25 03:22:26 +01:00
|
|
|
DistributionWorker.perform_async(status.id)
|
2016-11-08 01:32:34 +01:00
|
|
|
status
|
2016-03-25 02:13:30 +01:00
|
|
|
end
|
2016-02-28 21:22:56 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
def delete_status
|
|
|
|
Rails.logger.debug "Deleting remote status #{id}"
|
|
|
|
status = Status.find_by(uri: id)
|
|
|
|
RemoveStatusService.new.call(status) unless status.nil?
|
|
|
|
nil
|
|
|
|
end
|
2016-10-14 20:14:53 +02:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
def skip_unsupported_type?
|
|
|
|
!([:post, :share, :delete].include?(verb) && [:activity, :note, :comment].include?(type))
|
|
|
|
end
|
2016-02-28 21:22:56 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
def status_from_xml(entry)
|
|
|
|
# Return early if status already exists in db
|
|
|
|
status = find_status(id(entry))
|
|
|
|
return status unless status.nil?
|
|
|
|
|
|
|
|
status = Status.create!({
|
|
|
|
uri: id(entry),
|
|
|
|
url: url(entry),
|
|
|
|
account: account?(entry) ? find_or_resolve_account(acct(entry)) : @account,
|
|
|
|
text: content(entry),
|
|
|
|
created_at: published(entry),
|
|
|
|
})
|
|
|
|
|
|
|
|
if thread?(entry)
|
|
|
|
status.thread = find_or_resolve_status(status, *thread(entry))
|
|
|
|
end
|
2016-09-26 16:42:38 +02:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
mentions_from_xml(status, entry)
|
|
|
|
hashtags_from_xml(status, entry)
|
|
|
|
media_from_xml(status, entry)
|
2016-09-29 21:28:21 +02:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
status
|
|
|
|
end
|
2016-02-28 21:22:56 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
def find_or_resolve_account(acct)
|
|
|
|
FollowRemoteAccountService.new.call(acct)
|
|
|
|
end
|
2016-03-19 00:41:29 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
def find_or_resolve_status(parent, uri, url)
|
|
|
|
status = find_status(uri)
|
|
|
|
ThreadResolveWorker.perform_async(parent.id, url) if status.nil?
|
2016-09-22 21:10:36 +02:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
status
|
|
|
|
end
|
2016-09-26 16:42:38 +02:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
def find_status(uri)
|
|
|
|
if TagManager.instance.local_id?(uri)
|
|
|
|
local_id = TagManager.instance.unique_tag_to_local_id(uri, 'Status')
|
|
|
|
return Status.find(local_id)
|
2016-02-28 21:22:56 +01:00
|
|
|
end
|
2016-11-08 01:32:34 +01:00
|
|
|
|
|
|
|
Status.find_by(uri: uri)
|
2016-02-24 01:28:53 +01:00
|
|
|
end
|
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
def mentions_from_xml(parent, xml)
|
|
|
|
processed_account_ids = []
|
2016-10-14 20:14:53 +02:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
xml.xpath('./xmlns:link[@rel="mentioned"]').each do |link|
|
|
|
|
next if link['href'] == 'http://activityschema.org/collection/public'
|
2016-09-05 18:39:53 +02:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
url = Addressable::URI.parse(link['href'])
|
2016-09-22 20:42:20 +02:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
mentioned_account = if TagManager.instance.local_domain?(url.host)
|
|
|
|
Account.find_local(url.path.gsub('/users/', ''))
|
|
|
|
else
|
|
|
|
Account.find_by(url: link['href']) || FetchRemoteAccountService.new.call(link['href'])
|
|
|
|
end
|
2016-09-29 21:28:21 +02:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
next if mentioned_account.nil? || processed_account_ids.include?(mentioned_account.id)
|
2016-09-05 18:39:53 +02:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
if mentioned_account.local?
|
|
|
|
# Send notifications
|
|
|
|
NotificationMailer.mention(mentioned_account, parent).deliver_later unless mentioned_account.blocking?(parent.account)
|
|
|
|
end
|
2016-02-24 01:28:53 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
mentioned_account.mentions.where(status: parent).first_or_create(status: parent)
|
2016-02-24 17:23:59 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
# So we can skip duplicate mentions
|
|
|
|
processed_account_ids << mentioned_account.id
|
|
|
|
end
|
2016-03-19 19:20:07 +01:00
|
|
|
end
|
2016-02-24 01:28:53 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
def hashtags_from_xml(parent, xml)
|
|
|
|
tags = xml.xpath('./xmlns:category').map { |category| category['term'] }
|
|
|
|
ProcessHashtagsService.new.call(parent, tags)
|
Fix #24 - Thread resolving for remote statuses
This is a big one, so let me enumerate:
Accounts as well as stream entry pages now contain Link headers that
reference the Atom feed and Webfinger URL for the former and Atom entry
for the latter. So you only need to HEAD those resources to get that
information, no need to download and parse HTML <link>s.
ProcessFeedService will now queue ThreadResolveWorker for each remote
status that it cannot find otherwise. Furthermore, entries are now
processed in reverse order (from bottom to top) in case a newer entry
references a chronologically previous one.
ThreadResolveWorker uses FetchRemoteStatusService to obtain a status
and attach the child status it was queued for to it.
FetchRemoteStatusService looks up the URL, first with a HEAD, tests
if it's an Atom feed, in which case it processes it directly. Next
for Link headers to the Atom feed, in which case that is fetched
and processed. Lastly if it's HTML, it is checked for <link>s to the Atom
feed, and if such is found, that is fetched and processed. The account for
the status is derived from author/name attribute in the XML and the hostname
in the URL (domain). FollowRemoteAccountService and ProcessFeedService
are used.
This means that potentially threads are resolved recursively until a dead-end
is encountered, however it is performed asynchronously over background jobs,
so it should be ok.
2016-09-21 01:34:14 +02:00
|
|
|
end
|
2016-02-20 22:53:20 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
def media_from_xml(parent, xml)
|
|
|
|
xml.xpath('./xmlns:link[@rel="enclosure"]').each do |link|
|
|
|
|
next unless link['href']
|
2016-03-16 10:46:15 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
media = MediaAttachment.where(status: parent, remote_url: link['href']).first_or_initialize(account: parent.account, status: parent, remote_url: link['href'])
|
2016-02-20 22:53:20 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
begin
|
|
|
|
media.file_remote_url = link['href']
|
|
|
|
media.save
|
|
|
|
rescue Paperclip::Errors::NotIdentifiedByImageMagickError
|
|
|
|
next
|
|
|
|
end
|
|
|
|
end
|
2016-02-20 22:53:20 +01:00
|
|
|
end
|
2016-02-24 17:23:59 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
def id(xml = @xml)
|
|
|
|
xml.at_xpath('./xmlns:id').content
|
2016-02-24 17:23:59 +01:00
|
|
|
end
|
2016-02-24 01:28:53 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
def verb(xml = @xml)
|
|
|
|
raw = xml.at_xpath('./activity:verb', activity: ACTIVITY_NS).content
|
|
|
|
raw.gsub('http://activitystrea.ms/schema/1.0/', '').gsub('http://ostatus.org/schema/1.0/', '').to_sym
|
|
|
|
rescue
|
|
|
|
:post
|
Fix #24 - Thread resolving for remote statuses
This is a big one, so let me enumerate:
Accounts as well as stream entry pages now contain Link headers that
reference the Atom feed and Webfinger URL for the former and Atom entry
for the latter. So you only need to HEAD those resources to get that
information, no need to download and parse HTML <link>s.
ProcessFeedService will now queue ThreadResolveWorker for each remote
status that it cannot find otherwise. Furthermore, entries are now
processed in reverse order (from bottom to top) in case a newer entry
references a chronologically previous one.
ThreadResolveWorker uses FetchRemoteStatusService to obtain a status
and attach the child status it was queued for to it.
FetchRemoteStatusService looks up the URL, first with a HEAD, tests
if it's an Atom feed, in which case it processes it directly. Next
for Link headers to the Atom feed, in which case that is fetched
and processed. Lastly if it's HTML, it is checked for <link>s to the Atom
feed, and if such is found, that is fetched and processed. The account for
the status is derived from author/name attribute in the XML and the hostname
in the URL (domain). FollowRemoteAccountService and ProcessFeedService
are used.
This means that potentially threads are resolved recursively until a dead-end
is encountered, however it is performed asynchronously over background jobs,
so it should be ok.
2016-09-21 01:34:14 +02:00
|
|
|
end
|
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
def type(xml = @xml)
|
|
|
|
raw = xml.at_xpath('./activity:object-type', activity: ACTIVITY_NS).content
|
|
|
|
raw.gsub('http://activitystrea.ms/schema/1.0/', '').gsub('http://ostatus.org/schema/1.0/', '').to_sym
|
|
|
|
rescue
|
|
|
|
:activity
|
|
|
|
end
|
2016-02-24 01:28:53 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
def url(xml = @xml)
|
|
|
|
link = xml.at_xpath('./xmlns:link[@rel="alternate"]')
|
|
|
|
link['href']
|
|
|
|
end
|
2016-02-28 21:22:56 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
def content(xml = @xml)
|
|
|
|
xml.at_xpath('./xmlns:content').content
|
|
|
|
end
|
2016-02-24 17:23:59 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
def published(xml = @xml)
|
|
|
|
xml.at_xpath('./xmlns:published').content
|
|
|
|
end
|
2016-02-24 17:23:59 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
def thread?(xml = @xml)
|
|
|
|
!xml.at_xpath('./thr:in-reply-to', thr: THREAD_NS).nil?
|
|
|
|
end
|
2016-02-24 01:28:53 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
def thread(xml = @xml)
|
|
|
|
thr = xml.at_xpath('./thr:in-reply-to', thr: THREAD_NS)
|
|
|
|
[thr['ref'], thr['href']]
|
|
|
|
end
|
2016-02-24 03:05:40 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
def account?(xml = @xml)
|
|
|
|
!xml.at_xpath('./xmlns:author').nil?
|
|
|
|
end
|
2016-02-25 00:17:01 +01:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
def acct(xml = @xml)
|
|
|
|
username = xml.at_xpath('./xmlns:author/xmlns:name').content
|
|
|
|
url = xml.at_xpath('./xmlns:author/xmlns:uri').content
|
|
|
|
domain = Addressable::URI.parse(url).host
|
2016-09-20 00:39:03 +02:00
|
|
|
|
2016-11-08 01:32:34 +01:00
|
|
|
"#{username}@#{domain}"
|
|
|
|
end
|
2016-09-20 00:39:03 +02:00
|
|
|
end
|
2016-02-20 22:53:20 +01:00
|
|
|
end
|