forked from AkkomaGang/akkoma
some refactor and tests
This commit is contained in:
parent
503d966e9f
commit
f01ab6cd29
10 changed files with 308 additions and 201 deletions
|
@ -21,7 +21,10 @@
|
||||||
IO.puts(warning)
|
IO.puts(warning)
|
||||||
end
|
end
|
||||||
|
|
||||||
exported_config = Path.join([Path.dirname(config_path), "prod.exported_from_db.secret.exs"])
|
exported_config =
|
||||||
|
config_path
|
||||||
|
|> Path.dirname()
|
||||||
|
|> Path.join("prod.exported_from_db.secret.exs")
|
||||||
|
|
||||||
if File.exists?(exported_config) do
|
if File.exists?(exported_config) do
|
||||||
import_config exported_config
|
import_config exported_config
|
||||||
|
|
|
@ -18,11 +18,11 @@ mix pleroma.config migrate_to_db
|
||||||
|
|
||||||
## Transfer config from DB to `config/env.exported_from_db.secret.exs`
|
## Transfer config from DB to `config/env.exported_from_db.secret.exs`
|
||||||
|
|
||||||
|
To delete transfered settings from database optional flag `-d` can be used.
|
||||||
```sh tab="OTP"
|
```sh tab="OTP"
|
||||||
./bin/pleroma_ctl config migrate_from_db <env>
|
./bin/pleroma_ctl config migrate_from_db --env=<env> [-d]
|
||||||
```
|
```
|
||||||
|
|
||||||
```sh tab="From Source"
|
```sh tab="From Source"
|
||||||
mix pleroma.config migrate_from_db <env>
|
mix pleroma.config migrate_from_db --env=<env> [-d]
|
||||||
```
|
```
|
||||||
|
|
||||||
|
|
|
@ -19,94 +19,55 @@ def run(["migrate_to_db"]) do
|
||||||
end
|
end
|
||||||
|
|
||||||
def run(["migrate_from_db" | options]) do
|
def run(["migrate_from_db" | options]) do
|
||||||
# TODO: add support for releases
|
|
||||||
start_pleroma()
|
start_pleroma()
|
||||||
|
|
||||||
{opts, _} =
|
{opts, _} =
|
||||||
OptionParser.parse!(options,
|
OptionParser.parse!(options,
|
||||||
strict: [env: :string, delete_from_db: :boolean],
|
strict: [env: :string, delete: :boolean],
|
||||||
aliases: [d: :delete_from_db]
|
aliases: [d: :delete]
|
||||||
)
|
)
|
||||||
|
|
||||||
with {:active?, true} <-
|
migrate_from_db(opts)
|
||||||
{:active?, Pleroma.Config.get([:configurable_from_database])},
|
|
||||||
env when is_binary(env) <- opts[:env] || "prod",
|
|
||||||
config_path <- config_path(env),
|
|
||||||
{:ok, file} <- File.open(config_path, [:write, :utf8]) do
|
|
||||||
IO.write(file, config_header())
|
|
||||||
|
|
||||||
ConfigDB
|
|
||||||
|> Repo.all()
|
|
||||||
|> Enum.each(&write_to_file_with_deletion(&1, file, opts[:delete_from_db]))
|
|
||||||
|
|
||||||
File.close(file)
|
|
||||||
System.cmd("mix", ["format", config_path])
|
|
||||||
else
|
|
||||||
{:active?, false} ->
|
|
||||||
shell_info(
|
|
||||||
"Migration is not allowed by config. You can change this behavior in instance settings."
|
|
||||||
)
|
|
||||||
|
|
||||||
error ->
|
|
||||||
shell_info("Error occuried while opening file. #{inspect(error)}")
|
|
||||||
end
|
|
||||||
end
|
|
||||||
|
|
||||||
defp config_path(env) do
|
|
||||||
path =
|
|
||||||
if Pleroma.Config.get(:release) do
|
|
||||||
:config_path
|
|
||||||
|> Pleroma.Config.get()
|
|
||||||
|> Path.dirname()
|
|
||||||
else
|
|
||||||
"config"
|
|
||||||
end
|
|
||||||
|
|
||||||
Path.join(path, "#{env}.exported_from_db.secret.exs")
|
|
||||||
end
|
end
|
||||||
|
|
||||||
@spec migrate_to_db(Path.t() | nil) :: any()
|
@spec migrate_to_db(Path.t() | nil) :: any()
|
||||||
def migrate_to_db(file_path \\ nil) do
|
def migrate_to_db(file_path \\ nil) do
|
||||||
if Pleroma.Config.get([:configurable_from_database]) do
|
if Pleroma.Config.get([:configurable_from_database]) do
|
||||||
user_config_file =
|
config_file =
|
||||||
if Pleroma.Config.get(:release),
|
if file_path do
|
||||||
do: Pleroma.Config.get(:config_path),
|
file_path
|
||||||
else: "config/#{Pleroma.Config.get(:env)}.secret.exs"
|
else
|
||||||
|
if Pleroma.Config.get(:release) do
|
||||||
|
Pleroma.Config.get(:config_path)
|
||||||
|
else
|
||||||
|
"config/#{Pleroma.Config.get(:env)}.secret.exs"
|
||||||
|
end
|
||||||
|
end
|
||||||
|
|
||||||
config_file = file_path || user_config_file
|
|
||||||
do_migrate_to_db(config_file)
|
do_migrate_to_db(config_file)
|
||||||
else
|
else
|
||||||
shell_info(
|
migration_error()
|
||||||
"Migration is not allowed by config. You can change this behavior in instance settings."
|
|
||||||
)
|
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
if Code.ensure_loaded?(Config.Reader) do
|
|
||||||
defp config_header, do: "import Config\r\n\r\n"
|
|
||||||
defp read_file(config_file), do: Config.Reader.read_imports!(config_file)
|
|
||||||
else
|
|
||||||
defp config_header, do: "use Mix.Config\r\n\r\n"
|
|
||||||
defp read_file(config_file), do: Mix.Config.eval!(config_file)
|
|
||||||
end
|
|
||||||
|
|
||||||
defp do_migrate_to_db(config_file) do
|
defp do_migrate_to_db(config_file) do
|
||||||
if File.exists?(config_file) do
|
if File.exists?(config_file) do
|
||||||
{custom_config, _paths} = read_file(config_file)
|
custom_config =
|
||||||
|
config_file
|
||||||
|
|> read_file()
|
||||||
|
|> elem(0)
|
||||||
|
|
||||||
custom_config
|
custom_config
|
||||||
|> Keyword.keys()
|
|> Keyword.keys()
|
||||||
|> Enum.each(&create(&1, custom_config[&1]))
|
|> Enum.each(&create(&1, custom_config))
|
||||||
else
|
else
|
||||||
shell_info("To migrate settings, you must define custom settings in #{config_file}.")
|
shell_info("To migrate settings, you must define custom settings in #{config_file}.")
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
defp create(group, settings) do
|
defp create(group, settings) do
|
||||||
Enum.reject(settings, fn {k, _v} ->
|
group
|
||||||
k in [Pleroma.Repo, Pleroma.Web.Endpoint, :env, :configurable_from_database] or
|
|> Pleroma.Config.Loader.filter_group(settings)
|
||||||
(group == :phoenix and k == :serve_endpoints)
|
|
||||||
end)
|
|
||||||
|> Enum.each(fn {key, value} ->
|
|> Enum.each(fn {key, value} ->
|
||||||
key = inspect(key)
|
key = inspect(key)
|
||||||
{:ok, _} = ConfigDB.update_or_create(%{group: inspect(group), key: key, value: value})
|
{:ok, _} = ConfigDB.update_or_create(%{group: inspect(group), key: key, value: value})
|
||||||
|
@ -117,17 +78,70 @@ defp create(group, settings) do
|
||||||
shell_info("Settings for group :#{group} migrated.")
|
shell_info("Settings for group :#{group} migrated.")
|
||||||
end
|
end
|
||||||
|
|
||||||
defp write_to_file_with_deletion(config, file, with_deletion) do
|
defp migrate_from_db(opts) do
|
||||||
IO.write(
|
if Pleroma.Config.get([:configurable_from_database]) do
|
||||||
file,
|
env = opts[:env] || "prod"
|
||||||
"config #{config.group}, #{config.key}, #{
|
|
||||||
inspect(ConfigDB.from_binary(config.value), limit: :infinity)
|
|
||||||
}\r\n\r\n"
|
|
||||||
)
|
|
||||||
|
|
||||||
if with_deletion do
|
config_path =
|
||||||
|
if Pleroma.Config.get(:release) do
|
||||||
|
:config_path
|
||||||
|
|> Pleroma.Config.get()
|
||||||
|
|> Path.dirname()
|
||||||
|
else
|
||||||
|
"config"
|
||||||
|
end
|
||||||
|
|> Path.join("#{env}.exported_from_db.secret.exs")
|
||||||
|
|
||||||
|
file = File.open!(config_path, [:write, :utf8])
|
||||||
|
|
||||||
|
IO.write(file, config_header())
|
||||||
|
|
||||||
|
ConfigDB
|
||||||
|
|> Repo.all()
|
||||||
|
|> Enum.each(&write_and_delete(&1, file, opts[:delete]))
|
||||||
|
|
||||||
|
:ok = File.close(file)
|
||||||
|
System.cmd("mix", ["format", config_path])
|
||||||
|
else
|
||||||
|
migration_error()
|
||||||
|
end
|
||||||
|
end
|
||||||
|
|
||||||
|
defp migration_error do
|
||||||
|
shell_error(
|
||||||
|
"Migration is not allowed in config. You can change this behavior by setting `configurable_from_database` to true."
|
||||||
|
)
|
||||||
|
end
|
||||||
|
|
||||||
|
if Code.ensure_loaded?(Config.Reader) do
|
||||||
|
defp config_header, do: "import Config\r\n\r\n"
|
||||||
|
defp read_file(config_file), do: Config.Reader.read_imports!(config_file)
|
||||||
|
else
|
||||||
|
defp config_header, do: "use Mix.Config\r\n\r\n"
|
||||||
|
defp read_file(config_file), do: Mix.Config.eval!(config_file)
|
||||||
|
end
|
||||||
|
|
||||||
|
defp write_and_delete(config, file, delete?) do
|
||||||
|
config
|
||||||
|
|> write(file)
|
||||||
|
|> delete(delete?)
|
||||||
|
end
|
||||||
|
|
||||||
|
defp write(config, file) do
|
||||||
|
value =
|
||||||
|
config.value
|
||||||
|
|> ConfigDB.from_binary()
|
||||||
|
|> inspect(limit: :infinity)
|
||||||
|
|
||||||
|
IO.write(file, "config #{config.group}, #{config.key}, #{value}\r\n\r\n")
|
||||||
|
|
||||||
|
config
|
||||||
|
end
|
||||||
|
|
||||||
|
defp delete(config, true) do
|
||||||
{:ok, _} = Repo.delete(config)
|
{:ok, _} = Repo.delete(config)
|
||||||
shell_info("#{config.key} deleted from DB.")
|
shell_info("#{config.key} deleted from DB.")
|
||||||
end
|
end
|
||||||
end
|
|
||||||
|
defp delete(_config, _), do: :ok
|
||||||
end
|
end
|
||||||
|
|
|
@ -4,14 +4,38 @@
|
||||||
|
|
||||||
defmodule Pleroma.ConfigDB do
|
defmodule Pleroma.ConfigDB do
|
||||||
use Ecto.Schema
|
use Ecto.Schema
|
||||||
|
|
||||||
import Ecto.Changeset
|
import Ecto.Changeset
|
||||||
import Ecto.Query
|
import Ecto.Query
|
||||||
import Pleroma.Web.Gettext
|
import Pleroma.Web.Gettext
|
||||||
|
|
||||||
alias __MODULE__
|
alias __MODULE__
|
||||||
alias Pleroma.Repo
|
alias Pleroma.Repo
|
||||||
|
|
||||||
@type t :: %__MODULE__{}
|
@type t :: %__MODULE__{}
|
||||||
|
|
||||||
|
@full_key_update [
|
||||||
|
{:pleroma, :ecto_repos},
|
||||||
|
{:quack, :meta},
|
||||||
|
{:mime, :types},
|
||||||
|
{:cors_plug, [:max_age, :methods, :expose, :headers]},
|
||||||
|
{:auto_linker, :opts},
|
||||||
|
{:swarm, :node_blacklist},
|
||||||
|
{:logger, :backends}
|
||||||
|
]
|
||||||
|
|
||||||
|
@full_subkey_update [
|
||||||
|
{:pleroma, :assets, :mascots},
|
||||||
|
{:pleroma, :emoji, :groups},
|
||||||
|
{:pleroma, :workers, :retries},
|
||||||
|
{:pleroma, :mrf_subchain, :match_actor},
|
||||||
|
{:pleroma, :mrf_keyword, :replace}
|
||||||
|
]
|
||||||
|
|
||||||
|
@regex ~r/^~r(?'delimiter'[\/|"'([{<]{1})(?'pattern'.+)[\/|"')\]}>]{1}(?'modifier'[uismxfU]*)/u
|
||||||
|
|
||||||
|
@delimiters ["/", "|", "\"", "'", {"(", ")"}, {"[", "]"}, {"{", "}"}, {"<", ">"}]
|
||||||
|
|
||||||
schema "config" do
|
schema "config" do
|
||||||
field(:key, :string)
|
field(:key, :string)
|
||||||
field(:group, :string)
|
field(:group, :string)
|
||||||
|
@ -78,37 +102,35 @@ def get_db_keys(value, key) do
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
@full_subkey_update [
|
@spec merge_group(atom(), atom(), keyword(), keyword()) :: keyword()
|
||||||
{:pleroma, :assets, :mascots},
|
def merge_group(group, key, old_value, new_value) do
|
||||||
{:pleroma, :emoji, :groups},
|
new_keys = to_map_set(new_value)
|
||||||
{:pleroma, :workers, :retries},
|
|
||||||
{:pleroma, :mrf_subchain, :match_actor},
|
|
||||||
{:pleroma, :mrf_keyword, :replace}
|
|
||||||
]
|
|
||||||
|
|
||||||
@spec deep_merge(atom(), atom(), keyword(), keyword()) :: keyword()
|
intersect_keys =
|
||||||
def deep_merge(group, key, old_value, new_value) do
|
old_value |> to_map_set() |> MapSet.intersection(new_keys) |> MapSet.to_list()
|
||||||
old_keys =
|
|
||||||
old_value
|
|
||||||
|> Keyword.keys()
|
|
||||||
|> MapSet.new()
|
|
||||||
|
|
||||||
new_keys =
|
|
||||||
new_value
|
|
||||||
|> Keyword.keys()
|
|
||||||
|> MapSet.new()
|
|
||||||
|
|
||||||
intersect_keys = old_keys |> MapSet.intersection(new_keys) |> MapSet.to_list()
|
|
||||||
|
|
||||||
subkeys = sub_key_full_update(group, key, intersect_keys)
|
|
||||||
|
|
||||||
merged_value = ConfigDB.merge(old_value, new_value)
|
merged_value = ConfigDB.merge(old_value, new_value)
|
||||||
|
|
||||||
Enum.reduce(subkeys, merged_value, fn subkey, acc ->
|
@full_subkey_update
|
||||||
|
|> Enum.map(fn
|
||||||
|
{g, k, subkey} when g == group and k == key ->
|
||||||
|
if subkey in intersect_keys, do: subkey, else: []
|
||||||
|
|
||||||
|
_ ->
|
||||||
|
[]
|
||||||
|
end)
|
||||||
|
|> List.flatten()
|
||||||
|
|> Enum.reduce(merged_value, fn subkey, acc ->
|
||||||
Keyword.put(acc, subkey, new_value[subkey])
|
Keyword.put(acc, subkey, new_value[subkey])
|
||||||
end)
|
end)
|
||||||
end
|
end
|
||||||
|
|
||||||
|
defp to_map_set(keyword) do
|
||||||
|
keyword
|
||||||
|
|> Keyword.keys()
|
||||||
|
|> MapSet.new()
|
||||||
|
end
|
||||||
|
|
||||||
@spec sub_key_full_update?(atom(), atom(), [Keyword.key()]) :: boolean()
|
@spec sub_key_full_update?(atom(), atom(), [Keyword.key()]) :: boolean()
|
||||||
def sub_key_full_update?(group, key, subkeys) do
|
def sub_key_full_update?(group, key, subkeys) do
|
||||||
Enum.any?(@full_subkey_update, fn {g, k, subkey} ->
|
Enum.any?(@full_subkey_update, fn {g, k, subkey} ->
|
||||||
|
@ -116,17 +138,7 @@ def sub_key_full_update?(group, key, subkeys) do
|
||||||
end)
|
end)
|
||||||
end
|
end
|
||||||
|
|
||||||
defp sub_key_full_update(group, key, subkeys) do
|
@spec merge(keyword(), keyword()) :: keyword()
|
||||||
Enum.map(@full_subkey_update, fn
|
|
||||||
{g, k, subkey} when g == group and k == key ->
|
|
||||||
if subkey in subkeys, do: subkey, else: []
|
|
||||||
|
|
||||||
_ ->
|
|
||||||
[]
|
|
||||||
end)
|
|
||||||
|> List.flatten()
|
|
||||||
end
|
|
||||||
|
|
||||||
def merge(config1, config2) when is_list(config1) and is_list(config2) do
|
def merge(config1, config2) when is_list(config1) and is_list(config2) do
|
||||||
Keyword.merge(config1, config2, fn _, app1, app2 ->
|
Keyword.merge(config1, config2, fn _, app1, app2 ->
|
||||||
if Keyword.keyword?(app1) and Keyword.keyword?(app2) do
|
if Keyword.keyword?(app1) and Keyword.keyword?(app2) do
|
||||||
|
@ -145,31 +157,6 @@ defp deep_merge(_key, value1, value2) do
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
@full_key_update [
|
|
||||||
{:pleroma, :ecto_repos},
|
|
||||||
{:quack, :meta},
|
|
||||||
{:mime, :types},
|
|
||||||
{:cors_plug, [:max_age, :methods, :expose, :headers]},
|
|
||||||
{:auto_linker, :opts},
|
|
||||||
{:swarm, :node_blacklist},
|
|
||||||
{:logger, :backends}
|
|
||||||
]
|
|
||||||
|
|
||||||
defp only_full_update?(%ConfigDB{} = config) do
|
|
||||||
config_group = ConfigDB.from_string(config.group)
|
|
||||||
config_key = ConfigDB.from_string(config.key)
|
|
||||||
|
|
||||||
Enum.any?(@full_key_update, fn
|
|
||||||
{group, key} when is_list(key) ->
|
|
||||||
config_group == group and config_key in key
|
|
||||||
|
|
||||||
{group, key} ->
|
|
||||||
config_group == group and config_key == key
|
|
||||||
end)
|
|
||||||
end
|
|
||||||
|
|
||||||
defp can_be_partially_updated?(%ConfigDB{} = config), do: not only_full_update?(config)
|
|
||||||
|
|
||||||
@spec update_or_create(map()) :: {:ok, ConfigDB.t()} | {:error, Changeset.t()}
|
@spec update_or_create(map()) :: {:ok, ConfigDB.t()} | {:error, Changeset.t()}
|
||||||
def update_or_create(params) do
|
def update_or_create(params) do
|
||||||
search_opts = Map.take(params, [:group, :key])
|
search_opts = Map.take(params, [:group, :key])
|
||||||
|
@ -181,7 +168,7 @@ def update_or_create(params) do
|
||||||
transformed_value <- do_transform(params[:value]),
|
transformed_value <- do_transform(params[:value]),
|
||||||
{:can_be_merged, true, config} <- {:can_be_merged, is_list(transformed_value), config},
|
{:can_be_merged, true, config} <- {:can_be_merged, is_list(transformed_value), config},
|
||||||
new_value <-
|
new_value <-
|
||||||
deep_merge(
|
merge_group(
|
||||||
ConfigDB.from_string(config.group),
|
ConfigDB.from_string(config.group),
|
||||||
ConfigDB.from_string(config.key),
|
ConfigDB.from_string(config.key),
|
||||||
old_value,
|
old_value,
|
||||||
|
@ -197,6 +184,21 @@ def update_or_create(params) do
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
|
defp can_be_partially_updated?(%ConfigDB{} = config), do: not only_full_update?(config)
|
||||||
|
|
||||||
|
defp only_full_update?(%ConfigDB{} = config) do
|
||||||
|
config_group = ConfigDB.from_string(config.group)
|
||||||
|
config_key = ConfigDB.from_string(config.key)
|
||||||
|
|
||||||
|
Enum.any?(@full_key_update, fn
|
||||||
|
{group, key} when is_list(key) ->
|
||||||
|
config_group == group and config_key in key
|
||||||
|
|
||||||
|
{group, key} ->
|
||||||
|
config_group == group and config_key == key
|
||||||
|
end)
|
||||||
|
end
|
||||||
|
|
||||||
@spec delete(map()) :: {:ok, ConfigDB.t()} | {:error, Changeset.t()} | {:ok, nil}
|
@spec delete(map()) :: {:ok, ConfigDB.t()} | {:error, Changeset.t()} | {:ok, nil}
|
||||||
def delete(params) do
|
def delete(params) do
|
||||||
search_opts = Map.delete(params, :subkeys)
|
search_opts = Map.delete(params, :subkeys)
|
||||||
|
@ -286,18 +288,24 @@ defp do_convert({:proxy_url, {type, host, port}}) do
|
||||||
}
|
}
|
||||||
end
|
end
|
||||||
|
|
||||||
# TODO: will become useless after removing hackney
|
|
||||||
defp do_convert({:partial_chain, entity}), do: %{"tuple" => [":partial_chain", inspect(entity)]}
|
defp do_convert({:partial_chain, entity}), do: %{"tuple" => [":partial_chain", inspect(entity)]}
|
||||||
|
|
||||||
defp do_convert(entity) when is_tuple(entity),
|
defp do_convert(entity) when is_tuple(entity) do
|
||||||
do: %{"tuple" => do_convert(Tuple.to_list(entity))}
|
value =
|
||||||
|
entity
|
||||||
|
|> Tuple.to_list()
|
||||||
|
|> do_convert()
|
||||||
|
|
||||||
defp do_convert(entity) when is_boolean(entity) or is_number(entity) or is_nil(entity),
|
%{"tuple" => value}
|
||||||
do: entity
|
end
|
||||||
|
|
||||||
|
defp do_convert(entity) when is_boolean(entity) or is_number(entity) or is_nil(entity) do
|
||||||
|
entity
|
||||||
|
end
|
||||||
|
|
||||||
defp do_convert(entity)
|
defp do_convert(entity)
|
||||||
when is_atom(entity) and entity in [:"tlsv1.1", :"tlsv1.2", :"tlsv1.3"] do
|
when is_atom(entity) and entity in [:"tlsv1.1", :"tlsv1.2", :"tlsv1.3"] do
|
||||||
":#{to_string(entity)}"
|
":#{entity}"
|
||||||
end
|
end
|
||||||
|
|
||||||
defp do_convert(entity) when is_atom(entity), do: inspect(entity)
|
defp do_convert(entity) when is_atom(entity), do: inspect(entity)
|
||||||
|
@ -325,7 +333,6 @@ defp do_transform(%{"tuple" => [":proxy_url", %{"tuple" => [type, host, port]}]}
|
||||||
{:proxy_url, {do_transform_string(type), parse_host(host), port}}
|
{:proxy_url, {do_transform_string(type), parse_host(host), port}}
|
||||||
end
|
end
|
||||||
|
|
||||||
# TODO: will become useless after removing hackney
|
|
||||||
defp do_transform(%{"tuple" => [":partial_chain", entity]}) do
|
defp do_transform(%{"tuple" => [":partial_chain", entity]}) do
|
||||||
{partial_chain, []} =
|
{partial_chain, []} =
|
||||||
entity
|
entity
|
||||||
|
@ -369,10 +376,9 @@ defp parse_host(host) do
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
@delimiters ["/", "|", "\"", "'", {"(", ")"}, {"[", "]"}, {"{", "}"}, {"<", ">"}]
|
defp find_valid_delimiter([], _string, _) do
|
||||||
|
raise(ArgumentError, message: "valid delimiter for Regex expression not found")
|
||||||
defp find_valid_delimiter([], _string, _),
|
end
|
||||||
do: raise(ArgumentError, message: "valid delimiter for Regex expression not found")
|
|
||||||
|
|
||||||
defp find_valid_delimiter([{leading, closing} = delimiter | others], pattern, regex_delimiter)
|
defp find_valid_delimiter([{leading, closing} = delimiter | others], pattern, regex_delimiter)
|
||||||
when is_tuple(delimiter) do
|
when is_tuple(delimiter) do
|
||||||
|
@ -391,11 +397,9 @@ defp find_valid_delimiter([delimiter | others], pattern, regex_delimiter) do
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
@regex_parts ~r/^~r(?'delimiter'[\/|"'([{<]{1})(?'pattern'.+)[\/|"')\]}>]{1}(?'modifier'[uismxfU]*)/u
|
|
||||||
|
|
||||||
defp do_transform_string("~r" <> _pattern = regex) do
|
defp do_transform_string("~r" <> _pattern = regex) do
|
||||||
with %{"modifier" => modifier, "pattern" => pattern, "delimiter" => regex_delimiter} <-
|
with %{"modifier" => modifier, "pattern" => pattern, "delimiter" => regex_delimiter} <-
|
||||||
Regex.named_captures(@regex_parts, regex),
|
Regex.named_captures(@regex, regex),
|
||||||
{:ok, {leading, closing}} <- find_valid_delimiter(@delimiters, pattern, regex_delimiter),
|
{:ok, {leading, closing}} <- find_valid_delimiter(@delimiters, pattern, regex_delimiter),
|
||||||
{result, _} <- Code.eval_string("~r#{leading}#{pattern}#{closing}#{modifier}") do
|
{result, _} <- Code.eval_string("~r#{leading}#{pattern}#{closing}#{modifier}") do
|
||||||
result
|
result
|
||||||
|
|
|
@ -1,29 +1,16 @@
|
||||||
|
# Pleroma: A lightweight social networking server
|
||||||
|
# Copyright © 2017-2019 Pleroma Authors <https://pleroma.social/>
|
||||||
|
# SPDX-License-Identifier: AGPL-3.0-only
|
||||||
|
|
||||||
defmodule Pleroma.Config.Holder do
|
defmodule Pleroma.Config.Holder do
|
||||||
@config Pleroma.Config.Loader.load_and_merge()
|
@config Pleroma.Config.Loader.load_and_merge()
|
||||||
|
|
||||||
@spec config() :: keyword()
|
@spec config() :: keyword()
|
||||||
def config do
|
def config, do: @config
|
||||||
@config
|
|
||||||
|> Keyword.keys()
|
|
||||||
|> Enum.map(&filter(&1, config(&1)))
|
|
||||||
|> List.flatten()
|
|
||||||
end
|
|
||||||
|
|
||||||
@spec config(atom()) :: any()
|
@spec config(atom()) :: any()
|
||||||
def config(group), do: @config[group]
|
def config(group), do: @config[group]
|
||||||
|
|
||||||
@spec config(atom(), atom()) :: any()
|
@spec config(atom(), atom()) :: any()
|
||||||
def config(group, key), do: @config[group][key]
|
def config(group, key), do: @config[group][key]
|
||||||
|
|
||||||
defp filter(group, settings) when group not in [:swarm] do
|
|
||||||
filtered =
|
|
||||||
Enum.reject(settings, fn {k, _v} ->
|
|
||||||
k in [Pleroma.Repo, Pleroma.Web.Endpoint, :env, :configurable_from_database] or
|
|
||||||
(group == :phoenix and k == :serve_endpoints)
|
|
||||||
end)
|
|
||||||
|
|
||||||
{group, filtered}
|
|
||||||
end
|
|
||||||
|
|
||||||
defp filter(_, _), do: []
|
|
||||||
end
|
end
|
||||||
|
|
|
@ -1,6 +1,19 @@
|
||||||
|
# Pleroma: A lightweight social networking server
|
||||||
|
# Copyright © 2017-2019 Pleroma Authors <https://pleroma.social/>
|
||||||
|
# SPDX-License-Identifier: AGPL-3.0-only
|
||||||
|
|
||||||
defmodule Pleroma.Config.Loader do
|
defmodule Pleroma.Config.Loader do
|
||||||
@paths ["config/config.exs", "config/#{Mix.env()}.exs"]
|
@paths ["config/config.exs", "config/#{Mix.env()}.exs"]
|
||||||
|
|
||||||
|
@reject_keys [
|
||||||
|
Pleroma.Repo,
|
||||||
|
Pleroma.Web.Endpoint,
|
||||||
|
:env,
|
||||||
|
:configurable_from_database,
|
||||||
|
:database,
|
||||||
|
:swarm
|
||||||
|
]
|
||||||
|
|
||||||
if Code.ensure_loaded?(Config.Reader) do
|
if Code.ensure_loaded?(Config.Reader) do
|
||||||
@spec load(Path.t()) :: keyword()
|
@spec load(Path.t()) :: keyword()
|
||||||
def load(path), do: Config.Reader.read!(path)
|
def load(path), do: Config.Reader.read!(path)
|
||||||
|
@ -10,8 +23,9 @@ defp do_merge(conf1, conf2), do: Config.Reader.merge(conf1, conf2)
|
||||||
# support for Elixir less than 1.9
|
# support for Elixir less than 1.9
|
||||||
@spec load(Path.t()) :: keyword()
|
@spec load(Path.t()) :: keyword()
|
||||||
def load(path) do
|
def load(path) do
|
||||||
{config, _paths} = Mix.Config.eval!(path)
|
path
|
||||||
config
|
|> Mix.Config.eval!()
|
||||||
|
|> elem(0)
|
||||||
end
|
end
|
||||||
|
|
||||||
defp do_merge(conf1, conf2), do: Mix.Config.merge(conf1, conf2)
|
defp do_merge(conf1, conf2), do: Mix.Config.merge(conf1, conf2)
|
||||||
|
@ -26,14 +40,20 @@ def load_and_merge do
|
||||||
|
|
||||||
all_paths
|
all_paths
|
||||||
|> Enum.map(&load(&1))
|
|> Enum.map(&load(&1))
|
||||||
|> merge()
|
|> Enum.reduce([], &do_merge(&2, &1))
|
||||||
|
|> filter()
|
||||||
end
|
end
|
||||||
|
|
||||||
@spec merge([keyword()], keyword()) :: keyword()
|
defp filter(configs) do
|
||||||
def merge(configs, acc \\ [])
|
configs
|
||||||
def merge([], acc), do: acc
|
|> Keyword.keys()
|
||||||
|
|> Enum.reduce([], &Keyword.put(&2, &1, filter_group(&1, configs)))
|
||||||
|
end
|
||||||
|
|
||||||
def merge([config | others], acc) do
|
@spec filter_group(atom(), keyword()) :: keyword()
|
||||||
merge(others, do_merge(acc, config))
|
def filter_group(group, configs) do
|
||||||
|
Enum.reject(configs[group], fn {key, _v} ->
|
||||||
|
key in @reject_keys or (group == :phoenix and key == :serve_endpoints)
|
||||||
|
end)
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
|
@ -37,12 +37,11 @@ defp update_env(setting) do
|
||||||
group = ConfigDB.from_string(setting.group)
|
group = ConfigDB.from_string(setting.group)
|
||||||
value = ConfigDB.from_binary(setting.value)
|
value = ConfigDB.from_binary(setting.value)
|
||||||
|
|
||||||
if group != :phoenix and key != :serve_endpoints do
|
|
||||||
default = Pleroma.Config.Holder.config(group, key)
|
default = Pleroma.Config.Holder.config(group, key)
|
||||||
|
|
||||||
merged_value =
|
merged_value =
|
||||||
if can_be_merged?(default, value) do
|
if can_be_merged?(default, value) do
|
||||||
ConfigDB.deep_merge(group, key, default, value)
|
ConfigDB.merge_group(group, key, default, value)
|
||||||
else
|
else
|
||||||
value
|
value
|
||||||
end
|
end
|
||||||
|
@ -62,7 +61,6 @@ defp update_env(setting) do
|
||||||
|
|
||||||
nil
|
nil
|
||||||
end
|
end
|
||||||
end
|
|
||||||
rescue
|
rescue
|
||||||
e ->
|
e ->
|
||||||
Logger.warn(
|
Logger.warn(
|
||||||
|
@ -80,12 +78,15 @@ defp restart(started_applications, app) do
|
||||||
:ok <- Application.stop(app) do
|
:ok <- Application.stop(app) do
|
||||||
:ok = Application.start(app)
|
:ok = Application.start(app)
|
||||||
else
|
else
|
||||||
nil -> Logger.warn("#{app} is not started.")
|
nil ->
|
||||||
error -> Logger.warn(inspect(error))
|
Logger.warn("#{app} is not started.")
|
||||||
end
|
|
||||||
end
|
|
||||||
|
|
||||||
defp can_be_merged?(val1, val2) when is_map(val1) and is_map(val2), do: true
|
error ->
|
||||||
|
error
|
||||||
|
|> inspect()
|
||||||
|
|> Logger.warn()
|
||||||
|
end
|
||||||
|
end
|
||||||
|
|
||||||
defp can_be_merged?(val1, val2) when is_list(val1) and is_list(val2) do
|
defp can_be_merged?(val1, val2) when is_list(val1) and is_list(val2) do
|
||||||
Keyword.keyword?(val1) and Keyword.keyword?(val2)
|
Keyword.keyword?(val1) and Keyword.keyword?(val2)
|
||||||
|
|
|
@ -848,7 +848,7 @@ def config_show(conn, _params) do
|
||||||
merged_value =
|
merged_value =
|
||||||
if !is_nil(db_value) and Keyword.keyword?(db_value) and
|
if !is_nil(db_value) and Keyword.keyword?(db_value) and
|
||||||
ConfigDB.sub_key_full_update?(group, key, Keyword.keys(db_value)) do
|
ConfigDB.sub_key_full_update?(group, key, Keyword.keys(db_value)) do
|
||||||
ConfigDB.deep_merge(group, key, value, db_value)
|
ConfigDB.merge_group(group, key, value, db_value)
|
||||||
else
|
else
|
||||||
value
|
value
|
||||||
end
|
end
|
||||||
|
|
34
test/config/holder_test.exs
Normal file
34
test/config/holder_test.exs
Normal file
|
@ -0,0 +1,34 @@
|
||||||
|
# Pleroma: A lightweight social networking server
|
||||||
|
# Copyright © 2017-2019 Pleroma Authors <https://pleroma.social/>
|
||||||
|
# SPDX-License-Identifier: AGPL-3.0-only
|
||||||
|
|
||||||
|
defmodule Pleroma.Config.HolderTest do
|
||||||
|
use ExUnit.Case, async: true
|
||||||
|
|
||||||
|
alias Pleroma.Config.Holder
|
||||||
|
|
||||||
|
test "config/0" do
|
||||||
|
config = Holder.config()
|
||||||
|
assert config[:pleroma][Pleroma.Uploaders.Local][:uploads] == "test/uploads"
|
||||||
|
assert config[:tesla][:adapter] == Tesla.Mock
|
||||||
|
|
||||||
|
refute config[:pleroma][Pleroma.Repo]
|
||||||
|
refute config[:pleroma][Pleroma.Web.Endpoint]
|
||||||
|
refute config[:pleroma][:env]
|
||||||
|
refute config[:pleroma][:configurable_from_database]
|
||||||
|
refute config[:pleroma][:database]
|
||||||
|
refute config[:phoenix][:serve_endpoints]
|
||||||
|
end
|
||||||
|
|
||||||
|
test "config/1" do
|
||||||
|
pleroma_config = Holder.config(:pleroma)
|
||||||
|
assert pleroma_config[Pleroma.Uploaders.Local][:uploads] == "test/uploads"
|
||||||
|
tesla_config = Holder.config(:tesla)
|
||||||
|
assert tesla_config[:adapter] == Tesla.Mock
|
||||||
|
end
|
||||||
|
|
||||||
|
test "config/2" do
|
||||||
|
assert Holder.config(:pleroma, Pleroma.Uploaders.Local) == [uploads: "test/uploads"]
|
||||||
|
assert Holder.config(:tesla, :adapter) == Tesla.Mock
|
||||||
|
end
|
||||||
|
end
|
44
test/config/loader_test.exs
Normal file
44
test/config/loader_test.exs
Normal file
|
@ -0,0 +1,44 @@
|
||||||
|
# Pleroma: A lightweight social networking server
|
||||||
|
# Copyright © 2017-2019 Pleroma Authors <https://pleroma.social/>
|
||||||
|
# SPDX-License-Identifier: AGPL-3.0-only
|
||||||
|
|
||||||
|
defmodule Pleroma.Config.LoaderTest do
|
||||||
|
use ExUnit.Case, async: true
|
||||||
|
|
||||||
|
alias Pleroma.Config.Loader
|
||||||
|
|
||||||
|
test "load/1" do
|
||||||
|
config = Loader.load("test/fixtures/config/temp.secret.exs")
|
||||||
|
assert config[:pleroma][:first_setting][:key] == "value"
|
||||||
|
assert config[:pleroma][:first_setting][:key2] == [Pleroma.Repo]
|
||||||
|
assert config[:quack][:level] == :info
|
||||||
|
end
|
||||||
|
|
||||||
|
test "load_and_merge/0" do
|
||||||
|
config = Loader.load_and_merge()
|
||||||
|
|
||||||
|
refute config[:pleroma][Pleroma.Repo]
|
||||||
|
refute config[:pleroma][Pleroma.Web.Endpoint]
|
||||||
|
refute config[:pleroma][:env]
|
||||||
|
refute config[:pleroma][:configurable_from_database]
|
||||||
|
refute config[:pleroma][:database]
|
||||||
|
refute config[:phoenix][:serve_endpoints]
|
||||||
|
|
||||||
|
assert config[:pleroma][:ecto_repos] == [Pleroma.Repo]
|
||||||
|
assert config[:pleroma][Pleroma.Uploaders.Local][:uploads] == "test/uploads"
|
||||||
|
assert config[:tesla][:adapter] == Tesla.Mock
|
||||||
|
end
|
||||||
|
|
||||||
|
test "filter_group/2" do
|
||||||
|
assert Loader.filter_group(:pleroma,
|
||||||
|
pleroma: [
|
||||||
|
{Pleroma.Repo, [a: 1, b: 2]},
|
||||||
|
{Pleroma.Upload, [a: 1, b: 2]},
|
||||||
|
{Pleroma.Web.Endpoint, []},
|
||||||
|
env: :test,
|
||||||
|
configurable_from_database: true,
|
||||||
|
database: []
|
||||||
|
]
|
||||||
|
) == [{Pleroma.Upload, [a: 1, b: 2]}]
|
||||||
|
end
|
||||||
|
end
|
Loading…
Reference in a new issue