2018-12-23 20:04:54 +00:00
|
|
|
# Pleroma: A lightweight social networking server
|
2018-12-31 15:41:47 +00:00
|
|
|
# Copyright © 2017-2019 Pleroma Authors <https://pleroma.social/>
|
2018-12-23 20:04:54 +00:00
|
|
|
# SPDX-License-Identifier: AGPL-3.0-only
|
|
|
|
|
2017-03-21 08:21:52 +00:00
|
|
|
defmodule Pleroma.Object do
|
|
|
|
use Ecto.Schema
|
2019-02-09 15:16:26 +00:00
|
|
|
|
|
|
|
alias Pleroma.Activity
|
2019-03-05 02:52:23 +00:00
|
|
|
alias Pleroma.Object
|
2018-12-04 03:17:25 +00:00
|
|
|
alias Pleroma.Object.Fetcher
|
2019-02-09 15:16:26 +00:00
|
|
|
alias Pleroma.ObjectTombstone
|
2019-03-05 02:52:23 +00:00
|
|
|
alias Pleroma.Repo
|
|
|
|
alias Pleroma.User
|
2019-02-09 15:16:26 +00:00
|
|
|
|
|
|
|
import Ecto.Query
|
|
|
|
import Ecto.Changeset
|
2017-03-21 08:21:52 +00:00
|
|
|
|
2019-03-22 23:34:47 +00:00
|
|
|
require Logger
|
2017-03-21 08:21:52 +00:00
|
|
|
|
2020-01-12 18:48:58 +00:00
|
|
|
@type t() :: %__MODULE__{}
|
|
|
|
|
2017-03-21 08:21:52 +00:00
|
|
|
schema "objects" do
|
2018-03-30 13:01:53 +00:00
|
|
|
field(:data, :map)
|
2017-03-21 08:21:52 +00:00
|
|
|
|
|
|
|
timestamps()
|
|
|
|
end
|
2017-03-30 16:07:01 +00:00
|
|
|
|
2019-11-18 13:56:25 +00:00
|
|
|
def with_joined_activity(query, activity_type \\ "Create", join_type \\ :inner) do
|
|
|
|
object_position = Map.get(query.aliases, :object, 0)
|
|
|
|
|
|
|
|
join(query, join_type, [{object, object_position}], a in Activity,
|
|
|
|
on:
|
|
|
|
fragment(
|
|
|
|
"COALESCE(?->'object'->>'id', ?->>'object') = (? ->> 'id') AND (?->>'type' = ?) ",
|
|
|
|
a.data,
|
|
|
|
a.data,
|
|
|
|
object.data,
|
|
|
|
a.data,
|
|
|
|
^activity_type
|
|
|
|
),
|
|
|
|
as: :object_activity
|
|
|
|
)
|
|
|
|
end
|
|
|
|
|
2017-05-16 13:31:11 +00:00
|
|
|
def create(data) do
|
|
|
|
Object.change(%Object{}, %{data: data})
|
2018-03-30 13:01:53 +00:00
|
|
|
|> Repo.insert()
|
2017-05-16 13:31:11 +00:00
|
|
|
end
|
|
|
|
|
2017-05-09 16:11:51 +00:00
|
|
|
def change(struct, params \\ %{}) do
|
2017-11-19 01:22:07 +00:00
|
|
|
struct
|
2017-05-09 16:11:51 +00:00
|
|
|
|> cast(params, [:data])
|
|
|
|
|> validate_required([:data])
|
|
|
|
|> unique_constraint(:ap_id, name: :objects_unique_apid_index)
|
|
|
|
end
|
|
|
|
|
2019-05-21 17:40:35 +00:00
|
|
|
def get_by_id(nil), do: nil
|
|
|
|
def get_by_id(id), do: Repo.get(Object, id)
|
|
|
|
|
2019-09-18 15:13:21 +00:00
|
|
|
def get_by_id_and_maybe_refetch(id, opts \\ []) do
|
|
|
|
%{updated_at: updated_at} = object = get_by_id(id)
|
|
|
|
|
|
|
|
if opts[:interval] &&
|
2019-09-18 15:52:33 +00:00
|
|
|
NaiveDateTime.diff(NaiveDateTime.utc_now(), updated_at) > opts[:interval] do
|
2019-09-18 15:13:21 +00:00
|
|
|
case Fetcher.refetch_object(object) do
|
|
|
|
{:ok, %Object{} = object} ->
|
|
|
|
object
|
|
|
|
|
|
|
|
e ->
|
|
|
|
Logger.error("Couldn't refresh #{object.data["id"]}:\n#{inspect(e)}")
|
|
|
|
object
|
|
|
|
end
|
|
|
|
else
|
|
|
|
object
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2017-10-24 06:39:24 +00:00
|
|
|
def get_by_ap_id(nil), do: nil
|
2018-03-30 13:01:53 +00:00
|
|
|
|
2017-03-30 16:07:01 +00:00
|
|
|
def get_by_ap_id(ap_id) do
|
2018-03-30 13:01:53 +00:00
|
|
|
Repo.one(from(object in Object, where: fragment("(?)->>'id' = ?", object.data, ^ap_id)))
|
2017-03-30 16:07:01 +00:00
|
|
|
end
|
2017-04-30 11:53:26 +00:00
|
|
|
|
2020-01-12 18:48:58 +00:00
|
|
|
@doc """
|
|
|
|
Get a single attachment by it's name and href
|
|
|
|
"""
|
|
|
|
@spec get_attachment_by_name_and_href(String.t(), String.t()) :: Object.t() | nil
|
|
|
|
def get_attachment_by_name_and_href(name, href) do
|
|
|
|
query =
|
|
|
|
from(o in Object,
|
|
|
|
where: fragment("(?)->>'name' = ?", o.data, ^name),
|
|
|
|
where: fragment("(?)->>'href' = ?", o.data, ^href)
|
|
|
|
)
|
|
|
|
|
|
|
|
Repo.one(query)
|
|
|
|
end
|
|
|
|
|
2019-07-08 16:53:02 +00:00
|
|
|
defp warn_on_no_object_preloaded(ap_id) do
|
2019-11-21 13:03:01 +00:00
|
|
|
"Object.normalize() called without preloaded object (#{inspect(ap_id)}). Consider preloading the object"
|
2019-07-08 16:53:02 +00:00
|
|
|
|> Logger.debug()
|
|
|
|
|
|
|
|
Logger.debug("Backtrace: #{inspect(Process.info(:erlang.self(), :current_stacktrace))}")
|
|
|
|
end
|
|
|
|
|
2019-06-29 17:04:50 +00:00
|
|
|
def normalize(_, fetch_remote \\ true, options \\ [])
|
2019-07-08 16:53:02 +00:00
|
|
|
|
2019-03-22 23:34:47 +00:00
|
|
|
# If we pass an Activity to Object.normalize(), we can try to use the preloaded object.
|
|
|
|
# Use this whenever possible, especially when walking graphs in an O(N) loop!
|
2019-06-29 17:04:50 +00:00
|
|
|
def normalize(%Object{} = object, _, _), do: object
|
|
|
|
def normalize(%Activity{object: %Object{} = object}, _, _), do: object
|
2018-12-04 03:17:25 +00:00
|
|
|
|
2019-04-01 08:55:59 +00:00
|
|
|
# A hack for fake activities
|
2019-06-29 17:04:50 +00:00
|
|
|
def normalize(%Activity{data: %{"object" => %{"fake" => true} = data}}, _, _) do
|
2019-04-01 08:55:59 +00:00
|
|
|
%Object{id: "pleroma:fake_object_id", data: data}
|
|
|
|
end
|
|
|
|
|
2019-07-08 16:53:02 +00:00
|
|
|
# No preloaded object
|
2019-06-29 17:04:50 +00:00
|
|
|
def normalize(%Activity{data: %{"object" => %{"id" => ap_id}}}, fetch_remote, _) do
|
2019-07-08 16:53:02 +00:00
|
|
|
warn_on_no_object_preloaded(ap_id)
|
2019-04-17 09:22:32 +00:00
|
|
|
normalize(ap_id, fetch_remote)
|
2019-03-22 23:34:47 +00:00
|
|
|
end
|
|
|
|
|
2019-07-08 16:53:02 +00:00
|
|
|
# No preloaded object
|
2019-06-29 17:04:50 +00:00
|
|
|
def normalize(%Activity{data: %{"object" => ap_id}}, fetch_remote, _) do
|
2019-07-08 16:53:02 +00:00
|
|
|
warn_on_no_object_preloaded(ap_id)
|
2019-04-17 09:22:32 +00:00
|
|
|
normalize(ap_id, fetch_remote)
|
2019-03-22 23:34:47 +00:00
|
|
|
end
|
|
|
|
|
|
|
|
# Old way, try fetching the object through cache.
|
2019-06-29 17:04:50 +00:00
|
|
|
def normalize(%{"id" => ap_id}, fetch_remote, _), do: normalize(ap_id, fetch_remote)
|
|
|
|
def normalize(ap_id, false, _) when is_binary(ap_id), do: get_cached_by_ap_id(ap_id)
|
|
|
|
|
|
|
|
def normalize(ap_id, true, options) when is_binary(ap_id) do
|
|
|
|
Fetcher.fetch_object_from_id!(ap_id, options)
|
|
|
|
end
|
|
|
|
|
|
|
|
def normalize(_, _, _), do: nil
|
2018-06-18 05:23:54 +00:00
|
|
|
|
2018-12-06 07:26:17 +00:00
|
|
|
# Owned objects can only be mutated by their owner
|
|
|
|
def authorize_mutation(%Object{data: %{"actor" => actor}}, %User{ap_id: ap_id}),
|
|
|
|
do: actor == ap_id
|
|
|
|
|
|
|
|
# Legacy objects can be mutated by anybody
|
|
|
|
def authorize_mutation(%Object{}, %User{}), do: true
|
|
|
|
|
2019-02-03 17:28:14 +00:00
|
|
|
def get_cached_by_ap_id(ap_id) do
|
|
|
|
key = "object:#{ap_id}"
|
|
|
|
|
|
|
|
Cachex.fetch!(:object_cache, key, fn _ ->
|
|
|
|
object = get_by_ap_id(ap_id)
|
|
|
|
|
|
|
|
if object do
|
|
|
|
{:commit, object}
|
|
|
|
else
|
|
|
|
{:ignore, object}
|
|
|
|
end
|
|
|
|
end)
|
2017-05-01 14:12:20 +00:00
|
|
|
end
|
|
|
|
|
2017-04-30 11:53:26 +00:00
|
|
|
def context_mapping(context) do
|
2017-06-20 07:50:22 +00:00
|
|
|
Object.change(%Object{}, %{data: %{"id" => context}})
|
2017-04-30 11:53:26 +00:00
|
|
|
end
|
2018-11-01 07:28:48 +00:00
|
|
|
|
2018-12-25 00:00:06 +00:00
|
|
|
def make_tombstone(%Object{data: %{"id" => id, "type" => type}}, deleted \\ DateTime.utc_now()) do
|
|
|
|
%ObjectTombstone{
|
|
|
|
id: id,
|
|
|
|
formerType: type,
|
2018-12-23 23:25:36 +00:00
|
|
|
deleted: deleted
|
|
|
|
}
|
2018-12-25 00:00:06 +00:00
|
|
|
|> Map.from_struct()
|
2018-12-23 23:25:36 +00:00
|
|
|
end
|
|
|
|
|
2018-12-25 00:00:06 +00:00
|
|
|
def swap_object_with_tombstone(object) do
|
|
|
|
tombstone = make_tombstone(object)
|
2018-12-23 23:25:36 +00:00
|
|
|
|
|
|
|
object
|
|
|
|
|> Object.change(%{data: tombstone})
|
|
|
|
|> Repo.update()
|
|
|
|
end
|
|
|
|
|
2018-11-01 07:28:48 +00:00
|
|
|
def delete(%Object{data: %{"id" => id}} = object) do
|
2018-12-25 00:00:06 +00:00
|
|
|
with {:ok, _obj} = swap_object_with_tombstone(object),
|
2019-12-04 23:50:38 +00:00
|
|
|
deleted_activity = Activity.delete_all_by_object_ap_id(id),
|
2019-09-09 18:53:08 +00:00
|
|
|
{:ok, true} <- Cachex.del(:object_cache, "object:#{id}"),
|
2020-01-17 17:20:37 +00:00
|
|
|
{:ok, _} <- Cachex.del(:web_resp_cache, URI.parse(id).path),
|
|
|
|
:ok <- delete_attachments(object) do
|
2019-03-09 11:12:15 +00:00
|
|
|
{:ok, object, deleted_activity}
|
2018-11-01 07:28:48 +00:00
|
|
|
end
|
|
|
|
end
|
2019-02-03 17:28:14 +00:00
|
|
|
|
2020-01-12 18:48:58 +00:00
|
|
|
defp delete_attachments(%{data: %{"attachment" => [_ | _] = attachments, "actor" => actor}}) do
|
|
|
|
hrefs =
|
|
|
|
Enum.flat_map(attachments, fn attachment ->
|
|
|
|
Enum.map(attachment["url"], & &1["href"])
|
|
|
|
end)
|
|
|
|
|
|
|
|
names = Enum.map(attachments, & &1["name"])
|
|
|
|
|
|
|
|
uploader = Pleroma.Config.get([Pleroma.Upload, :uploader])
|
|
|
|
|
|
|
|
# find all objects for copies of the attachments, name and actor doesn't matter here
|
|
|
|
delete_ids =
|
|
|
|
from(o in Object,
|
|
|
|
where:
|
|
|
|
fragment(
|
2020-01-14 16:53:28 +00:00
|
|
|
"to_jsonb(array(select jsonb_array_elements((?)#>'{url}') ->> 'href' where jsonb_typeof((?)#>'{url}') = 'array'))::jsonb \\?| (?)",
|
|
|
|
o.data,
|
2020-01-12 18:48:58 +00:00
|
|
|
o.data,
|
|
|
|
^hrefs
|
|
|
|
)
|
|
|
|
)
|
|
|
|
|> Repo.all()
|
|
|
|
# we should delete 1 object for any given attachment, but don't delete files if
|
|
|
|
# there are more than 1 object for it
|
|
|
|
|> Enum.reduce(%{}, fn %{
|
|
|
|
id: id,
|
|
|
|
data: %{
|
|
|
|
"url" => [%{"href" => href}],
|
|
|
|
"actor" => obj_actor,
|
|
|
|
"name" => name
|
|
|
|
}
|
|
|
|
},
|
|
|
|
acc ->
|
|
|
|
Map.update(acc, href, %{id: id, count: 1}, fn val ->
|
|
|
|
case obj_actor == actor and name in names do
|
|
|
|
true ->
|
|
|
|
# set id of the actor's object that will be deleted
|
|
|
|
%{val | id: id, count: val.count + 1}
|
|
|
|
|
|
|
|
false ->
|
|
|
|
# another actor's object, just increase count to not delete file
|
|
|
|
%{val | count: val.count + 1}
|
|
|
|
end
|
|
|
|
end)
|
|
|
|
end)
|
|
|
|
|> Enum.map(fn {href, %{id: id, count: count}} ->
|
|
|
|
# only delete files that have single instance
|
|
|
|
with 1 <- count do
|
|
|
|
prefix =
|
|
|
|
case Pleroma.Config.get([Pleroma.Upload, :base_url]) do
|
|
|
|
nil -> "media"
|
|
|
|
_ -> ""
|
|
|
|
end
|
|
|
|
|
|
|
|
base_url = Pleroma.Config.get([__MODULE__, :base_url], Pleroma.Web.base_url())
|
|
|
|
|
|
|
|
file_path = String.trim_leading(href, "#{base_url}/#{prefix}")
|
|
|
|
|
|
|
|
uploader.delete_file(file_path)
|
|
|
|
end
|
|
|
|
|
|
|
|
id
|
|
|
|
end)
|
|
|
|
|
|
|
|
from(o in Object, where: o.id in ^delete_ids)
|
|
|
|
|> Repo.delete_all()
|
|
|
|
|
|
|
|
:ok
|
|
|
|
end
|
|
|
|
|
|
|
|
defp delete_attachments(%{data: _data}), do: :ok
|
|
|
|
|
2019-05-21 00:41:40 +00:00
|
|
|
def prune(%Object{data: %{"id" => id}} = object) do
|
|
|
|
with {:ok, object} <- Repo.delete(object),
|
2019-09-09 18:53:08 +00:00
|
|
|
{:ok, true} <- Cachex.del(:object_cache, "object:#{id}"),
|
|
|
|
{:ok, _} <- Cachex.del(:web_resp_cache, URI.parse(id).path) do
|
2019-05-21 00:41:40 +00:00
|
|
|
{:ok, object}
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2019-02-03 17:28:14 +00:00
|
|
|
def set_cache(%Object{data: %{"id" => ap_id}} = object) do
|
|
|
|
Cachex.put(:object_cache, "object:#{ap_id}", object)
|
|
|
|
{:ok, object}
|
|
|
|
end
|
|
|
|
|
|
|
|
def update_and_set_cache(changeset) do
|
|
|
|
with {:ok, object} <- Repo.update(changeset) do
|
|
|
|
set_cache(object)
|
|
|
|
end
|
|
|
|
end
|
2019-03-25 17:21:48 +00:00
|
|
|
|
|
|
|
def increase_replies_count(ap_id) do
|
|
|
|
Object
|
|
|
|
|> where([o], fragment("?->>'id' = ?::text", o.data, ^to_string(ap_id)))
|
|
|
|
|> update([o],
|
|
|
|
set: [
|
|
|
|
data:
|
|
|
|
fragment(
|
|
|
|
"""
|
2019-10-18 11:11:30 +00:00
|
|
|
safe_jsonb_set(?, '{repliesCount}',
|
2019-03-25 17:21:48 +00:00
|
|
|
(coalesce((?->>'repliesCount')::int, 0) + 1)::varchar::jsonb, true)
|
|
|
|
""",
|
|
|
|
o.data,
|
|
|
|
o.data
|
|
|
|
)
|
|
|
|
]
|
|
|
|
)
|
|
|
|
|> Repo.update_all([])
|
|
|
|
|> case do
|
|
|
|
{1, [object]} -> set_cache(object)
|
|
|
|
_ -> {:error, "Not found"}
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
|
|
|
def decrease_replies_count(ap_id) do
|
|
|
|
Object
|
|
|
|
|> where([o], fragment("?->>'id' = ?::text", o.data, ^to_string(ap_id)))
|
|
|
|
|> update([o],
|
|
|
|
set: [
|
|
|
|
data:
|
|
|
|
fragment(
|
|
|
|
"""
|
2019-10-18 11:11:30 +00:00
|
|
|
safe_jsonb_set(?, '{repliesCount}',
|
2019-03-25 17:21:48 +00:00
|
|
|
(greatest(0, (?->>'repliesCount')::int - 1))::varchar::jsonb, true)
|
|
|
|
""",
|
|
|
|
o.data,
|
|
|
|
o.data
|
|
|
|
)
|
|
|
|
]
|
|
|
|
)
|
|
|
|
|> Repo.update_all([])
|
|
|
|
|> case do
|
|
|
|
{1, [object]} -> set_cache(object)
|
|
|
|
_ -> {:error, "Not found"}
|
2018-11-01 07:28:48 +00:00
|
|
|
end
|
|
|
|
end
|
2019-05-21 11:12:10 +00:00
|
|
|
|
|
|
|
def increase_vote_count(ap_id, name) do
|
|
|
|
with %Object{} = object <- Object.normalize(ap_id),
|
|
|
|
"Question" <- object.data["type"] do
|
|
|
|
multiple = Map.has_key?(object.data, "anyOf")
|
|
|
|
|
|
|
|
options =
|
|
|
|
(object.data["anyOf"] || object.data["oneOf"] || [])
|
|
|
|
|> Enum.map(fn
|
|
|
|
%{"name" => ^name} = option ->
|
|
|
|
Kernel.update_in(option["replies"]["totalItems"], &(&1 + 1))
|
|
|
|
|
|
|
|
option ->
|
|
|
|
option
|
|
|
|
end)
|
|
|
|
|
|
|
|
data =
|
|
|
|
if multiple do
|
|
|
|
Map.put(object.data, "anyOf", options)
|
|
|
|
else
|
|
|
|
Map.put(object.data, "oneOf", options)
|
|
|
|
end
|
|
|
|
|
|
|
|
object
|
|
|
|
|> Object.change(%{data: data})
|
|
|
|
|> update_and_set_cache()
|
|
|
|
else
|
|
|
|
_ -> :noop
|
|
|
|
end
|
|
|
|
end
|
2019-09-06 18:50:00 +00:00
|
|
|
|
|
|
|
@doc "Updates data field of an object"
|
|
|
|
def update_data(%Object{data: data} = object, attrs \\ %{}) do
|
|
|
|
object
|
|
|
|
|> Object.change(%{data: Map.merge(data || %{}, attrs)})
|
|
|
|
|> Repo.update()
|
|
|
|
end
|
2019-11-23 19:55:41 +00:00
|
|
|
|
|
|
|
def local?(%Object{data: %{"id" => id}}) do
|
|
|
|
String.starts_with?(id, Pleroma.Web.base_url() <> "/")
|
|
|
|
end
|
2017-03-21 08:21:52 +00:00
|
|
|
end
|