2019-07-10 05:13:23 +00:00
|
|
|
# Pleroma: A lightweight social networking server
|
2021-01-13 06:49:20 +00:00
|
|
|
# Copyright © 2017-2021 Pleroma Authors <https://pleroma.social/>
|
2019-07-10 05:13:23 +00:00
|
|
|
# SPDX-License-Identifier: AGPL-3.0-only
|
|
|
|
|
2018-12-01 22:53:10 +00:00
|
|
|
defmodule Pleroma.Object.Fetcher do
|
2019-05-25 04:24:21 +00:00
|
|
|
alias Pleroma.HTTP
|
2022-11-15 17:23:47 +00:00
|
|
|
alias Pleroma.Instances
|
2020-09-15 15:22:08 +00:00
|
|
|
alias Pleroma.Maps
|
2019-04-17 09:22:32 +00:00
|
|
|
alias Pleroma.Object
|
2018-12-01 22:53:10 +00:00
|
|
|
alias Pleroma.Object.Containment
|
2019-09-18 15:13:21 +00:00
|
|
|
alias Pleroma.Repo
|
2019-07-17 22:41:42 +00:00
|
|
|
alias Pleroma.Signature
|
|
|
|
alias Pleroma.Web.ActivityPub.InternalFetchActor
|
2020-06-14 20:01:14 +00:00
|
|
|
alias Pleroma.Web.ActivityPub.ObjectValidator
|
2018-12-01 22:53:10 +00:00
|
|
|
alias Pleroma.Web.ActivityPub.Transmogrifier
|
2020-02-15 17:41:38 +00:00
|
|
|
alias Pleroma.Web.Federator
|
2018-12-01 22:53:10 +00:00
|
|
|
|
|
|
|
require Logger
|
2019-09-18 16:53:51 +00:00
|
|
|
require Pleroma.Constants
|
2018-12-01 22:53:10 +00:00
|
|
|
|
2024-03-13 20:42:51 +00:00
|
|
|
@moduledoc """
|
|
|
|
This module deals with correctly fetching Acitivity Pub objects in a safe way.
|
|
|
|
|
|
|
|
The core function is `fetch_and_contain_remote_object_from_id/1` which performs
|
|
|
|
the actual fetch and common safety and authenticity checks. Other `fetch_*`
|
|
|
|
function use the former and perform some additional tasks
|
|
|
|
"""
|
|
|
|
|
2024-03-13 21:12:17 +00:00
|
|
|
@mix_env Mix.env()
|
|
|
|
|
2019-09-18 16:07:25 +00:00
|
|
|
defp touch_changeset(changeset) do
|
|
|
|
updated_at =
|
|
|
|
NaiveDateTime.utc_now()
|
|
|
|
|> NaiveDateTime.truncate(:second)
|
|
|
|
|
|
|
|
Ecto.Changeset.put_change(changeset, :updated_at, updated_at)
|
|
|
|
end
|
|
|
|
|
2020-06-14 20:24:00 +00:00
|
|
|
defp maybe_reinject_internal_fields(%{data: %{} = old_data}, new_data) do
|
2022-09-06 19:24:02 +00:00
|
|
|
has_history? = fn
|
|
|
|
%{"formerRepresentations" => %{"orderedItems" => list}} when is_list(list) -> true
|
|
|
|
_ -> false
|
|
|
|
end
|
|
|
|
|
2019-09-18 16:53:51 +00:00
|
|
|
internal_fields = Map.take(old_data, Pleroma.Constants.object_internal_fields())
|
|
|
|
|
2022-09-06 19:24:02 +00:00
|
|
|
remote_history_exists? = has_history?.(new_data)
|
|
|
|
|
|
|
|
# If the remote history exists, we treat that as the only source of truth.
|
|
|
|
new_data =
|
|
|
|
if has_history?.(old_data) and not remote_history_exists? do
|
|
|
|
Map.put(new_data, "formerRepresentations", old_data["formerRepresentations"])
|
|
|
|
else
|
|
|
|
new_data
|
|
|
|
end
|
|
|
|
|
|
|
|
# If the remote does not have history information, we need to manage it ourselves
|
|
|
|
new_data =
|
|
|
|
if not remote_history_exists? do
|
|
|
|
changed? =
|
|
|
|
Pleroma.Constants.status_updatable_fields()
|
|
|
|
|> Enum.any?(fn field -> Map.get(old_data, field) != Map.get(new_data, field) end)
|
|
|
|
|
|
|
|
%{updated_object: updated_object} =
|
|
|
|
new_data
|
|
|
|
|> Object.Updater.maybe_update_history(old_data,
|
|
|
|
updated: changed?,
|
|
|
|
use_history_in_new_object?: false
|
|
|
|
)
|
|
|
|
|
|
|
|
updated_object
|
|
|
|
else
|
|
|
|
new_data
|
|
|
|
end
|
|
|
|
|
2020-06-14 20:24:00 +00:00
|
|
|
Map.merge(new_data, internal_fields)
|
2019-09-18 16:53:51 +00:00
|
|
|
end
|
|
|
|
|
2020-06-14 20:24:00 +00:00
|
|
|
defp maybe_reinject_internal_fields(_, new_data), do: new_data
|
2019-09-18 16:53:51 +00:00
|
|
|
|
2019-09-19 04:35:34 +00:00
|
|
|
@spec reinject_object(struct(), map()) :: {:ok, Object.t()} | {:error, any()}
|
2020-06-14 20:01:14 +00:00
|
|
|
defp reinject_object(%Object{data: %{"type" => "Question"}} = object, new_data) do
|
|
|
|
Logger.debug("Reinjecting object #{new_data["id"]}")
|
|
|
|
|
2020-08-25 00:21:19 +00:00
|
|
|
with data <- maybe_reinject_internal_fields(object, new_data),
|
2020-06-14 20:01:14 +00:00
|
|
|
{:ok, data, _} <- ObjectValidator.validate(data, %{}),
|
|
|
|
changeset <- Object.change(object, %{data: data}),
|
|
|
|
changeset <- touch_changeset(changeset),
|
|
|
|
{:ok, object} <- Repo.insert_or_update(changeset),
|
|
|
|
{:ok, object} <- Object.set_cache(object) do
|
|
|
|
{:ok, object}
|
|
|
|
else
|
|
|
|
e ->
|
|
|
|
Logger.error("Error while processing object: #{inspect(e)}")
|
|
|
|
{:error, e}
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2020-06-14 20:24:00 +00:00
|
|
|
defp reinject_object(%Object{} = object, new_data) do
|
|
|
|
Logger.debug("Reinjecting object #{new_data["id"]}")
|
2019-05-21 00:41:58 +00:00
|
|
|
|
2020-06-14 20:24:00 +00:00
|
|
|
with new_data <- Transmogrifier.fix_object(new_data),
|
|
|
|
data <- maybe_reinject_internal_fields(object, new_data),
|
|
|
|
changeset <- Object.change(object, %{data: data}),
|
2019-09-18 16:07:25 +00:00
|
|
|
changeset <- touch_changeset(changeset),
|
2019-11-06 11:00:03 +00:00
|
|
|
{:ok, object} <- Repo.insert_or_update(changeset),
|
|
|
|
{:ok, object} <- Object.set_cache(object) do
|
2019-05-21 00:41:58 +00:00
|
|
|
{:ok, object}
|
|
|
|
else
|
|
|
|
e ->
|
|
|
|
Logger.error("Error while processing object: #{inspect(e)}")
|
|
|
|
{:error, e}
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2024-03-13 20:42:51 +00:00
|
|
|
@doc "Assumes object already is in our database and refetches from remote to update (e.g. for polls)"
|
2019-09-18 15:13:21 +00:00
|
|
|
def refetch_object(%Object{data: %{"id" => id}} = object) do
|
2019-11-23 19:55:41 +00:00
|
|
|
with {:local, false} <- {:local, Object.local?(object)},
|
2024-07-13 04:54:37 +00:00
|
|
|
{:ok, new_data} <- fetch_and_contain_remote_object_from_id(id, true),
|
2024-03-24 18:32:28 +00:00
|
|
|
{:id, true} <- {:id, new_data["id"] == id},
|
2020-06-14 20:24:00 +00:00
|
|
|
{:ok, object} <- reinject_object(object, new_data) do
|
2019-09-18 15:13:21 +00:00
|
|
|
{:ok, object}
|
|
|
|
else
|
2019-11-06 22:40:55 +00:00
|
|
|
{:local, true} -> {:ok, object}
|
2024-04-13 21:56:04 +00:00
|
|
|
{:id, false} -> {:error, :id_mismatch}
|
2019-09-18 15:13:21 +00:00
|
|
|
e -> {:error, e}
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2024-03-13 20:42:51 +00:00
|
|
|
@doc """
|
|
|
|
Fetches a new object and puts it through the processing pipeline for inbound objects
|
|
|
|
|
|
|
|
Note: will also insert a fake Create activity, since atm we internally
|
|
|
|
need everything to be traced back to a Create activity.
|
|
|
|
"""
|
2019-06-29 17:04:50 +00:00
|
|
|
def fetch_object_from_id(id, options \\ []) do
|
2022-12-10 00:09:45 +00:00
|
|
|
with %URI{} = uri <- URI.parse(id),
|
2023-08-04 11:50:50 +00:00
|
|
|
# let's check the URI is even vaguely valid first
|
2024-04-13 22:55:26 +00:00
|
|
|
{:valid_uri_scheme, true} <-
|
|
|
|
{:valid_uri_scheme, uri.scheme == "http" or uri.scheme == "https"},
|
2022-12-10 00:09:45 +00:00
|
|
|
# If we have instance restrictions, apply them here to prevent fetching from unwanted instances
|
2024-04-13 22:55:26 +00:00
|
|
|
{:mrf_reject_check, {:ok, nil}} <-
|
|
|
|
{:mrf_reject_check, Pleroma.Web.ActivityPub.MRF.SimplePolicy.check_reject(uri)},
|
|
|
|
{:mrf_accept_check, {:ok, _}} <-
|
|
|
|
{:mrf_accept_check, Pleroma.Web.ActivityPub.MRF.SimplePolicy.check_accept(uri)},
|
2022-12-10 00:09:45 +00:00
|
|
|
{_, nil} <- {:fetch_object, Object.get_cached_by_ap_id(id)},
|
2020-02-15 17:41:38 +00:00
|
|
|
{_, true} <- {:allowed_depth, Federator.allowed_thread_distance?(options[:depth])},
|
|
|
|
{_, {:ok, data}} <- {:fetch, fetch_and_contain_remote_object_from_id(id)},
|
2021-01-04 12:38:31 +00:00
|
|
|
{_, nil} <- {:normalize, Object.normalize(data, fetch: false)},
|
2019-09-11 04:23:33 +00:00
|
|
|
params <- prepare_activity_params(data),
|
2020-02-15 17:41:38 +00:00
|
|
|
{_, {:ok, activity}} <-
|
2019-10-18 04:08:25 +00:00
|
|
|
{:transmogrifier, Transmogrifier.handle_incoming(params, options)},
|
2020-02-15 17:41:38 +00:00
|
|
|
{_, _data, %Object{} = object} <-
|
2021-01-04 12:38:31 +00:00
|
|
|
{:object, data, Object.normalize(activity, fetch: false)} do
|
2018-12-01 22:53:10 +00:00
|
|
|
{:ok, object}
|
|
|
|
else
|
2023-12-28 01:47:18 +00:00
|
|
|
{:allowed_depth, false} = e ->
|
|
|
|
log_fetch_error(id, e)
|
2023-12-28 03:28:41 +00:00
|
|
|
{:error, :allowed_depth}
|
2020-02-15 17:41:38 +00:00
|
|
|
|
2024-04-13 21:56:04 +00:00
|
|
|
{:valid_uri_scheme, _} = e ->
|
|
|
|
log_fetch_error(id, e)
|
|
|
|
{:error, :invalid_uri_scheme}
|
2023-08-04 11:50:50 +00:00
|
|
|
|
2024-04-13 22:55:26 +00:00
|
|
|
{:mrf_reject_check, _} = e ->
|
|
|
|
log_fetch_error(id, e)
|
|
|
|
{:reject, :mrf}
|
|
|
|
|
|
|
|
{:mrf_accept_check, _} = e ->
|
|
|
|
log_fetch_error(id, e)
|
|
|
|
{:reject, :mrf}
|
|
|
|
|
2023-12-28 01:47:18 +00:00
|
|
|
{:containment, reason} = e ->
|
|
|
|
log_fetch_error(id, e)
|
|
|
|
{:error, reason}
|
2023-12-26 19:22:04 +00:00
|
|
|
|
2023-12-28 01:47:18 +00:00
|
|
|
{:transmogrifier, {:error, {:reject, reason}}} = e ->
|
|
|
|
log_fetch_error(id, e)
|
|
|
|
{:reject, reason}
|
2018-12-01 22:53:10 +00:00
|
|
|
|
2023-12-28 01:47:18 +00:00
|
|
|
{:transmogrifier, {:reject, reason}} = e ->
|
|
|
|
log_fetch_error(id, e)
|
|
|
|
{:reject, reason}
|
2020-09-17 14:17:16 +00:00
|
|
|
|
2023-12-28 01:47:18 +00:00
|
|
|
{:transmogrifier, reason} = e ->
|
|
|
|
log_fetch_error(id, e)
|
|
|
|
{:error, reason}
|
2019-10-18 03:41:38 +00:00
|
|
|
|
2019-09-11 04:23:33 +00:00
|
|
|
{:object, data, nil} ->
|
2019-09-19 04:35:34 +00:00
|
|
|
reinject_object(%Object{}, data)
|
2019-05-21 00:41:58 +00:00
|
|
|
|
2019-09-11 04:23:33 +00:00
|
|
|
{:normalize, object = %Object{}} ->
|
|
|
|
{:ok, object}
|
2018-12-01 22:53:10 +00:00
|
|
|
|
2019-09-11 04:23:33 +00:00
|
|
|
{:fetch_object, %Object{} = object} ->
|
|
|
|
{:ok, object}
|
2018-12-01 22:53:10 +00:00
|
|
|
|
2023-12-28 01:47:18 +00:00
|
|
|
{:fetch, {:error, reason}} = e ->
|
|
|
|
log_fetch_error(id, e)
|
|
|
|
{:error, reason}
|
2019-10-24 16:08:34 +00:00
|
|
|
|
2019-10-17 23:37:21 +00:00
|
|
|
e ->
|
2023-12-28 01:47:18 +00:00
|
|
|
log_fetch_error(id, e)
|
2023-12-26 19:20:33 +00:00
|
|
|
{:error, e}
|
2018-12-01 22:53:10 +00:00
|
|
|
end
|
2019-09-11 04:23:33 +00:00
|
|
|
end
|
|
|
|
|
2023-12-28 01:47:18 +00:00
|
|
|
defp log_fetch_error(id, error) do
|
2023-12-28 02:55:07 +00:00
|
|
|
Logger.metadata(object: id)
|
2023-12-28 01:47:18 +00:00
|
|
|
Logger.error("Object rejected while fetching #{id} #{inspect(error)}")
|
|
|
|
end
|
|
|
|
|
2019-09-11 04:23:33 +00:00
|
|
|
defp prepare_activity_params(data) do
|
|
|
|
%{
|
|
|
|
"type" => "Create",
|
|
|
|
# Should we seriously keep this attributedTo thing?
|
|
|
|
"actor" => data["actor"] || data["attributedTo"],
|
|
|
|
"object" => data
|
|
|
|
}
|
2020-09-15 15:22:08 +00:00
|
|
|
|> Maps.put_if_present("to", data["to"])
|
|
|
|
|> Maps.put_if_present("cc", data["cc"])
|
|
|
|
|> Maps.put_if_present("bto", data["bto"])
|
|
|
|
|> Maps.put_if_present("bcc", data["bcc"])
|
2018-12-01 22:53:10 +00:00
|
|
|
end
|
|
|
|
|
2019-07-17 22:41:42 +00:00
|
|
|
defp make_signature(id, date) do
|
|
|
|
uri = URI.parse(id)
|
|
|
|
|
|
|
|
signature =
|
|
|
|
InternalFetchActor.get_actor()
|
|
|
|
|> Signature.sign(%{
|
|
|
|
"(request-target)": "get #{uri.path}",
|
|
|
|
host: uri.host,
|
|
|
|
date: date
|
|
|
|
})
|
|
|
|
|
2020-09-02 06:16:51 +00:00
|
|
|
{"signature", signature}
|
2019-07-17 22:41:42 +00:00
|
|
|
end
|
|
|
|
|
|
|
|
defp sign_fetch(headers, id, date) do
|
|
|
|
if Pleroma.Config.get([:activitypub, :sign_object_fetches]) do
|
2020-09-02 06:16:51 +00:00
|
|
|
[make_signature(id, date) | headers]
|
2019-07-17 22:41:42 +00:00
|
|
|
else
|
|
|
|
headers
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
defp maybe_date_fetch(headers, date) do
|
|
|
|
if Pleroma.Config.get([:activitypub, :sign_object_fetches]) do
|
2020-09-02 06:16:51 +00:00
|
|
|
[{"date", date} | headers]
|
2019-07-17 22:41:42 +00:00
|
|
|
else
|
|
|
|
headers
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2024-07-13 04:54:37 +00:00
|
|
|
@doc """
|
|
|
|
Fetches arbitrary remote object and performs basic safety and authenticity checks.
|
|
|
|
When the fetch URL is known to already be a canonical AP id, checks are stricter.
|
|
|
|
"""
|
|
|
|
def fetch_and_contain_remote_object_from_id(id, is_ap_id \\ false)
|
2020-09-18 11:58:22 +00:00
|
|
|
|
2024-07-13 04:54:37 +00:00
|
|
|
def fetch_and_contain_remote_object_from_id(%{"id" => id}, is_ap_id),
|
|
|
|
do: fetch_and_contain_remote_object_from_id(id, is_ap_id)
|
2020-09-18 11:58:22 +00:00
|
|
|
|
2024-07-13 04:54:37 +00:00
|
|
|
def fetch_and_contain_remote_object_from_id(id, is_ap_id) when is_binary(id) do
|
|
|
|
Logger.debug("Fetching object #{id} via AP [ap_id=#{is_ap_id}]")
|
2018-12-01 22:53:10 +00:00
|
|
|
|
2024-04-13 21:56:04 +00:00
|
|
|
with {:valid_uri_scheme, true} <- {:valid_uri_scheme, String.starts_with?(id, "http")},
|
2024-04-13 22:55:26 +00:00
|
|
|
%URI{} = uri <- URI.parse(id),
|
|
|
|
{:mrf_reject_check, {:ok, nil}} <-
|
|
|
|
{:mrf_reject_check, Pleroma.Web.ActivityPub.MRF.SimplePolicy.check_reject(uri)},
|
|
|
|
{:mrf_accept_check, {:ok, _}} <-
|
|
|
|
{:mrf_accept_check, Pleroma.Web.ActivityPub.MRF.SimplePolicy.check_accept(uri)},
|
|
|
|
{:local_fetch, :ok} <- {:local_fetch, Containment.contain_local_fetch(id)},
|
2024-03-15 19:57:09 +00:00
|
|
|
{:ok, final_id, body} <- get_object(id),
|
2024-07-13 04:54:37 +00:00
|
|
|
# a canonical ID shouldn't be a redirect
|
|
|
|
true <- !is_ap_id || final_id == id,
|
2020-09-18 11:58:22 +00:00
|
|
|
{:ok, data} <- safe_json_decode(body),
|
2024-07-13 04:54:37 +00:00
|
|
|
{_, :ok} <- {:containment, Containment.contain_origin(final_id, data)},
|
|
|
|
{_, _, :ok} <- {:strict_id, data["id"], Containment.contain_id_to_fetch(final_id, data)} do
|
2024-03-15 19:57:09 +00:00
|
|
|
unless Instances.reachable?(final_id) do
|
|
|
|
Instances.set_reachable(final_id)
|
2022-11-15 17:23:47 +00:00
|
|
|
end
|
|
|
|
|
2020-09-18 11:58:22 +00:00
|
|
|
{:ok, data}
|
|
|
|
else
|
2024-07-13 04:54:37 +00:00
|
|
|
# E.g. Mastodon and *key serve the AP object directly under their display URLs without
|
|
|
|
# redirecting to their canonical location first, thus ids will expectedly differ.
|
|
|
|
# Similarly keys, either use a fragment ID and are a subobjects or a distinct ID
|
|
|
|
# but for compatibility are still a subobject presenting their owning actors ID at the toplevel.
|
|
|
|
# Refetching _once_ from the listed id, should yield a strict match afterwards.
|
|
|
|
{:strict_id, ap_id, _} = e ->
|
|
|
|
case is_ap_id do
|
|
|
|
false ->
|
|
|
|
fetch_and_contain_remote_object_from_id(ap_id, true)
|
|
|
|
|
|
|
|
true ->
|
|
|
|
log_fetch_error(id, e)
|
|
|
|
{:error, :id_mismatch}
|
|
|
|
end
|
Only allow exact id matches
This protects us from falling for obvious spoofs as from the current
upload exploit (unfortunately we can’t reasonably do anything about
spoofs with exact matches as was possible via emoji and proxy).
Such objects being invalid is supported by the spec, sepcifically
sections 3.1 and 3.2: https://www.w3.org/TR/activitypub/#obj-id
Anonymous objects are not relevant here (they can only exists within
parent objects iiuc) and neither is client-to-server or transient objects
(as those cannot be fetched in the first place).
This leaves us with the requirement for `id` to (a) exist and
(b) be a publicly dereferencable URI from the originating server.
This alone does not yet demand strict equivalence, but the spec then
further explains objects ought to be fetchable _via their ID_.
Meaning an object not retrievable via its ID, is invalid.
This reading is supported by the fact, e.g. GoToSocial (recently) and
Mastodon (for 6+ years) do already implement such strict ID checks,
additionally proving this doesn’t cause federation issues in practice.
However, apart from canonical IDs there can also be additional display
URLs. *omas first redirect those to their canonical location, but *keys
and Mastodon directly serve the AP representation without redirects.
Mastodon and GTS deal with this in two different ways,
but both constitute an effective countermeasure:
- Mastodon:
Unless it already is a known AP id, two fetches occur.
The first fetch just reads the `id` property and then refetches from
the id. The last fetch requires the returned id to exactly match the
URL the content was fetched from. (This can be optimised by skipping
the second fetch if it already matches)
https://github.com/mastodon/mastodon/blob/05eda8d19330a9c27c0cf07de19a87edff269057/app/helpers/jsonld_helper.rb#L168
https://github.com/mastodon/mastodon/commit/63f097979990bf5ba9db848b8a253056bad781af
- GTS:
Only does a single fetch and then checks if _either_ the id
_or_ url property (which can be an object) match the original fetch
URL. This relies on implementations always including their display URL
as "url" if differing from the id. For actors this is true for all
investigated implementations, for posts only Mastodon includes an
"url", but it is also the only one with a differing display URL.
https://github.com/superseriousbusiness/gotosocial/commit/2bafd7daf542d985ee76d9079a30a602cb7be827#diff-943bbb02c8ac74ac5dc5d20807e561dcdfaebdc3b62b10730f643a20ac23c24fR222
Albeit Mastodon’s refetch offers higher compatibility with theoretical
implmentations using either multiple different display URL or not
denoting any of them as "url" at all, for now we chose to adopt a
GTS-like refetch-free approach to avoid additional implementation
concerns wrt to whether redirects should be allowed when fetching a
canonical AP id and potential for accidentally loosening some checks
(e.g. cross-domain refetches) for one of the fetches.
This may be reconsidered in the future.
2024-03-16 00:00:19 +00:00
|
|
|
|
2024-04-13 22:55:26 +00:00
|
|
|
{:mrf_reject_check, _} = e ->
|
|
|
|
log_fetch_error(id, e)
|
|
|
|
{:reject, :mrf}
|
|
|
|
|
|
|
|
{:mrf_accept_check, _} = e ->
|
|
|
|
log_fetch_error(id, e)
|
|
|
|
{:reject, :mrf}
|
|
|
|
|
2024-04-13 21:56:04 +00:00
|
|
|
{:valid_uri_scheme, _} = e ->
|
|
|
|
log_fetch_error(id, e)
|
|
|
|
{:error, :invalid_uri_scheme}
|
2020-09-18 11:58:22 +00:00
|
|
|
|
2024-04-13 21:56:04 +00:00
|
|
|
{:local_fetch, _} = e ->
|
|
|
|
log_fetch_error(id, e)
|
|
|
|
{:error, :local_resource}
|
2024-03-13 22:00:23 +00:00
|
|
|
|
2024-04-13 21:56:04 +00:00
|
|
|
{:containment, reason} ->
|
|
|
|
log_fetch_error(id, reason)
|
2024-07-13 04:54:45 +00:00
|
|
|
{:error, {:containment, reason}}
|
2024-03-13 21:21:19 +00:00
|
|
|
|
2020-09-18 11:58:22 +00:00
|
|
|
{:error, e} ->
|
|
|
|
{:error, e}
|
|
|
|
|
|
|
|
e ->
|
|
|
|
{:error, e}
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2024-07-13 04:54:37 +00:00
|
|
|
def fetch_and_contain_remote_object_from_id(_id, _is_ap_id),
|
2024-04-13 21:56:04 +00:00
|
|
|
do: {:error, :invalid_id}
|
2020-09-18 11:58:22 +00:00
|
|
|
|
2024-03-13 21:12:17 +00:00
|
|
|
# HOPEFULLY TEMPORARY
|
|
|
|
# Basically none of our Tesla mocks in tests set the (supposed to
|
|
|
|
# exist for Tesla proper) url parameter for their responses
|
|
|
|
# causing almost every fetch in test to fail otherwise
|
2024-03-15 19:57:09 +00:00
|
|
|
if @mix_env == :test do
|
|
|
|
defp get_final_id(nil, initial_url), do: initial_url
|
|
|
|
defp get_final_id("", initial_url), do: initial_url
|
|
|
|
end
|
|
|
|
|
|
|
|
defp get_final_id(final_url, _intial_url) do
|
|
|
|
final_url
|
|
|
|
end
|
|
|
|
|
2022-12-12 19:06:04 +00:00
|
|
|
def get_object(id) do
|
2019-08-22 19:39:06 +00:00
|
|
|
date = Pleroma.Signature.signed_date()
|
2019-07-17 22:41:42 +00:00
|
|
|
|
|
|
|
headers =
|
2024-04-11 20:45:45 +00:00
|
|
|
[
|
|
|
|
# The first is required by spec, the second provided as a fallback for buggy implementations
|
|
|
|
{"accept", "application/ld+json; profile=\"https://www.w3.org/ns/activitystreams\""},
|
|
|
|
{"accept", "application/activity+json"}
|
|
|
|
]
|
2019-07-17 22:41:42 +00:00
|
|
|
|> maybe_date_fetch(date)
|
|
|
|
|> sign_fetch(id, date)
|
|
|
|
|
2024-03-13 21:12:17 +00:00
|
|
|
with {:ok, %{body: body, status: code, headers: headers, url: final_url}}
|
|
|
|
when code in 200..299 <-
|
2023-12-15 17:12:45 +00:00
|
|
|
HTTP.Backoff.get(id, headers),
|
2024-03-11 21:52:46 +00:00
|
|
|
{:has_content_type, {_, content_type}} <-
|
|
|
|
{:has_content_type, List.keyfind(headers, "content-type", 0)},
|
|
|
|
{:parse_content_type, {:ok, "application", subtype, type_params}} <-
|
|
|
|
{:parse_content_type, Plug.Conn.Utils.media_type(content_type)} do
|
2024-03-15 19:57:09 +00:00
|
|
|
final_id = get_final_id(final_url, id)
|
|
|
|
|
2024-03-11 21:52:46 +00:00
|
|
|
case {subtype, type_params} do
|
|
|
|
{"activity+json", _} ->
|
2024-03-15 19:57:09 +00:00
|
|
|
{:ok, final_id, body}
|
2024-03-11 21:52:46 +00:00
|
|
|
|
|
|
|
{"ld+json", %{"profile" => "https://www.w3.org/ns/activitystreams"}} ->
|
2024-03-15 19:57:09 +00:00
|
|
|
{:ok, final_id, body}
|
2024-03-11 21:52:46 +00:00
|
|
|
|
|
|
|
_ ->
|
|
|
|
{:error, {:content_type, content_type}}
|
|
|
|
end
|
|
|
|
else
|
2023-12-29 04:09:33 +00:00
|
|
|
{:ok, %{status: code}} when code in [401, 403] ->
|
2023-12-28 02:57:47 +00:00
|
|
|
{:error, :forbidden}
|
2023-12-26 21:05:28 +00:00
|
|
|
|
2019-06-13 10:13:35 +00:00
|
|
|
{:ok, %{status: code}} when code in [404, 410] ->
|
2023-12-28 02:57:47 +00:00
|
|
|
{:error, :not_found}
|
2019-06-13 09:34:03 +00:00
|
|
|
|
2019-10-24 16:08:34 +00:00
|
|
|
{:error, e} ->
|
|
|
|
{:error, e}
|
|
|
|
|
2024-03-11 21:52:46 +00:00
|
|
|
{:has_content_type, _} ->
|
|
|
|
{:error, {:content_type, nil}}
|
|
|
|
|
|
|
|
{:parse_content_type, e} ->
|
|
|
|
{:error, {:content_type, e}}
|
|
|
|
|
2019-10-18 02:42:25 +00:00
|
|
|
e ->
|
|
|
|
{:error, e}
|
2018-12-01 22:53:10 +00:00
|
|
|
end
|
|
|
|
end
|
2019-07-13 16:17:57 +00:00
|
|
|
|
2020-09-18 11:58:22 +00:00
|
|
|
defp safe_json_decode(nil), do: {:ok, nil}
|
|
|
|
defp safe_json_decode(json), do: Jason.decode(json)
|
2018-12-01 22:53:10 +00:00
|
|
|
end
|