if Code.ensure_loaded?(Postgrex) do defmodule Ecto.Adapters.Postgres.Connection do @moduledoc false @default_port 5432 @behaviour Ecto.Adapters.SQL.Connection @explain_prepared_statement_name "ecto_explain_statement" ## Module and Options @impl true def child_spec(opts) do opts |> Keyword.put_new(:port, @default_port) |> Postgrex.child_spec() end @impl true def to_constraints( %Postgrex.Error{postgres: %{code: :unique_violation, constraint: constraint}}, _opts ), do: [unique: constraint] def to_constraints( %Postgrex.Error{postgres: %{code: :foreign_key_violation, constraint: constraint}}, _opts ), do: [foreign_key: constraint] def to_constraints( %Postgrex.Error{postgres: %{code: :exclusion_violation, constraint: constraint}}, _opts ), do: [exclusion: constraint] def to_constraints( %Postgrex.Error{postgres: %{code: :check_violation, constraint: constraint}}, _opts ), do: [check: constraint] # Postgres 9.2 and earlier does not provide the constraint field @impl true def to_constraints( %Postgrex.Error{postgres: %{code: :unique_violation, message: message}}, _opts ) do case :binary.split(message, " unique constraint ") do [_, quoted] -> [unique: strip_quotes(quoted)] _ -> [] end end def to_constraints( %Postgrex.Error{postgres: %{code: :foreign_key_violation, message: message}}, _opts ) do case :binary.split(message, " foreign key constraint ") do [_, quoted] -> [quoted | _] = :binary.split(quoted, " on table ") [foreign_key: strip_quotes(quoted)] _ -> [] end end def to_constraints( %Postgrex.Error{postgres: %{code: :exclusion_violation, message: message}}, _opts ) do case :binary.split(message, " exclusion constraint ") do [_, quoted] -> [exclusion: strip_quotes(quoted)] _ -> [] end end def to_constraints( %Postgrex.Error{postgres: %{code: :check_violation, message: message}}, _opts ) do case :binary.split(message, " check constraint ") do [_, quoted] -> [check: strip_quotes(quoted)] _ -> [] end end def to_constraints(_, _opts), do: [] defp strip_quotes(quoted) do size = byte_size(quoted) - 2 <<_, unquoted::binary-size(size), _>> = quoted unquoted end ## Query @impl true def prepare_execute(conn, name, sql, params, opts) do ensure_list_params!(params) case Postgrex.prepare_execute(conn, name, sql, params, opts) do {:error, %Postgrex.Error{postgres: %{pg_code: "22P02", message: message}} = error} -> context = """ . If you are trying to query a JSON field, the parameter may need to be interpolated. \ Instead of p.json["field"] != "value" do p.json["field"] != ^"value" """ {:error, put_in(error.postgres.message, message <> context)} other -> other end end @impl true def query(conn, sql, params, opts) do ensure_list_params!(params) Postgrex.query(conn, sql, params, opts) end @impl true def query_many(_conn, _sql, _params, _opts) do raise RuntimeError, "query_many is not supported in the PostgreSQL adapter" end @impl true def execute(conn, %{ref: ref} = query, params, opts) do ensure_list_params!(params) case Postgrex.execute(conn, query, params, opts) do {:ok, %{ref: ^ref}, result} -> {:ok, result} {:ok, _, _} = ok -> ok {:error, %Postgrex.QueryError{} = err} -> {:reset, err} {:error, %Postgrex.Error{postgres: %{code: :feature_not_supported}} = err} -> {:reset, err} {:error, _} = error -> error end end @impl true def stream(conn, sql, params, opts) do ensure_list_params!(params) Postgrex.stream(conn, sql, params, opts) end defp ensure_list_params!(params) do unless is_list(params) do raise ArgumentError, "expected params to be a list, got: #{inspect(params)}" end end @parent_as __MODULE__ alias Ecto.Query.{BooleanExpr, ByExpr, JoinExpr, QueryExpr, WithExpr} @impl true def all(query, as_prefix \\ []) do sources = create_names(query, as_prefix) {select_distinct, order_by_distinct} = distinct(query.distinct, sources, query) cte = cte(query, sources) from = from(query, sources) select = select(query, select_distinct, sources) join = join(query, sources) where = where(query, sources) group_by = group_by(query, sources) having = having(query, sources) window = window(query, sources) combinations = combinations(query, as_prefix) order_by = order_by(query, order_by_distinct, sources) limit = limit(query, sources) offset = offset(query, sources) lock = lock(query, sources) [ cte, select, from, join, where, group_by, having, window, combinations, order_by, limit, offset | lock ] end @impl true def update_all(%{from: %{source: source}} = query, prefix \\ nil) do sources = create_names(query, []) cte = cte(query, sources) {from, name} = get_source(query, sources, 0, source) prefix = prefix || ["UPDATE ", from, " AS ", name | " SET "] fields = update_fields(query, sources) {join, wheres} = using_join(query, :update_all, "FROM", sources) where = where(%{query | wheres: wheres ++ query.wheres}, sources) [cte, prefix, fields, join, where | returning(query, sources)] end @impl true def delete_all(%{from: from} = query) do sources = create_names(query, []) cte = cte(query, sources) {from, name} = get_source(query, sources, 0, from) {join, wheres} = using_join(query, :delete_all, "USING", sources) where = where(%{query | wheres: wheres ++ query.wheres}, sources) [cte, "DELETE FROM ", from, " AS ", name, join, where | returning(query, sources)] end @impl true def insert(prefix, table, header, rows, on_conflict, returning, placeholders) do counter_offset = length(placeholders) + 1 values = if header == [] do [" VALUES " | Enum.map_intersperse(rows, ?,, fn _ -> "(DEFAULT)" end)] else [" (", quote_names(header), ") " | insert_all(rows, counter_offset)] end [ "INSERT INTO ", quote_name(prefix, table), insert_as(on_conflict), values, on_conflict(on_conflict, header) | returning(returning) ] end defp insert_as({%{sources: sources}, _, _}) do {_expr, name, _schema} = create_name(sources, 0, []) [" AS " | name] end defp insert_as({_, _, _}) do [] end defp on_conflict({:raise, _, []}, _header), do: [] defp on_conflict({:nothing, _, targets}, _header), do: [" ON CONFLICT ", conflict_target(targets) | "DO NOTHING"] defp on_conflict({fields, _, targets}, _header) when is_list(fields), do: [" ON CONFLICT ", conflict_target!(targets), "DO " | replace(fields)] defp on_conflict({query, _, targets}, _header), do: [" ON CONFLICT ", conflict_target!(targets), "DO " | update_all(query, "UPDATE SET ")] defp conflict_target!([]), do: error!(nil, "the :conflict_target option is required on upserts by PostgreSQL") defp conflict_target!(target), do: conflict_target(target) defp conflict_target({:unsafe_fragment, fragment}), do: [fragment, ?\s] defp conflict_target([]), do: [] defp conflict_target(targets), do: [?(, quote_names(targets), ?), ?\s] defp replace(fields) do [ "UPDATE SET " | Enum.map_intersperse(fields, ?,, fn field -> quoted = quote_name(field) [quoted, " = ", "EXCLUDED." | quoted] end) ] end defp insert_all(query = %Ecto.Query{}, _counter) do [?(, all(query), ?)] end defp insert_all(rows, counter) do [ "VALUES ", intersperse_reduce(rows, ?,, counter, fn row, counter -> {row, counter} = insert_each(row, counter) {[?(, row, ?)], counter} end) |> elem(0) ] end defp insert_each(values, counter) do intersperse_reduce(values, ?,, counter, fn nil, counter -> {"DEFAULT", counter} {%Ecto.Query{} = query, params_counter}, counter -> {[?(, all(query), ?)], counter + params_counter} {:placeholder, placeholder_index}, counter -> {[?$ | placeholder_index], counter} _, counter -> {[?$ | Integer.to_string(counter)], counter + 1} end) end @impl true def update(prefix, table, fields, filters, returning) do {fields, count} = intersperse_reduce(fields, ", ", 1, fn field, acc -> {[quote_name(field), " = $" | Integer.to_string(acc)], acc + 1} end) {filters, _count} = intersperse_reduce(filters, " AND ", count, fn {field, nil}, acc -> {[quote_name(field), " IS NULL"], acc} {field, _value}, acc -> {[quote_name(field), " = $" | Integer.to_string(acc)], acc + 1} end) [ "UPDATE ", quote_name(prefix, table), " SET ", fields, " WHERE ", filters | returning(returning) ] end @impl true def delete(prefix, table, filters, returning) do {filters, _} = intersperse_reduce(filters, " AND ", 1, fn {field, nil}, acc -> {[quote_name(field), " IS NULL"], acc} {field, _value}, acc -> {[quote_name(field), " = $" | Integer.to_string(acc)], acc + 1} end) ["DELETE FROM ", quote_name(prefix, table), " WHERE ", filters | returning(returning)] end @impl true def explain_query(conn, query, params, opts) do {explain_opts, opts} = Keyword.split( opts, ~w[analyze verbose costs settings buffers timing summary format plan]a ) {plan_type, explain_opts} = Keyword.pop(explain_opts, :plan) fallback_generic? = plan_type == :fallback_generic result = cond do fallback_generic? and explain_opts[:analyze] -> raise ArgumentError, "analyze cannot be used with a `:fallback_generic` explain plan " <> "as the actual parameter values are ignored under this plan type." <> "You may either change the plan type to `:custom` or remove the `:analyze` option." fallback_generic? -> explain_queries = build_fallback_generic_queries(query, length(params), explain_opts) fallback_generic_query(conn, explain_queries, opts) true -> query(conn, build_explain_query(query, explain_opts), params, opts) end map_format? = explain_opts[:format] == :map case result do {:ok, %Postgrex.Result{rows: rows}} when map_format? -> {:ok, List.flatten(rows)} {:ok, %Postgrex.Result{rows: rows}} -> {:ok, Enum.map_join(rows, "\n", & &1)} error -> error end end def build_fallback_generic_queries(query, num_params, opts) do prepare_args = if num_params > 0, do: ["( ", Enum.map_intersperse(1..num_params, ", ", fn _ -> "unknown" end), " )"], else: [] prepare = [ "PREPARE ", @explain_prepared_statement_name, prepare_args, " AS ", query ] |> IO.iodata_to_binary() set = "SET LOCAL plan_cache_mode = force_generic_plan" execute_args = if num_params > 0, do: ["( ", Enum.map_intersperse(1..num_params, ", ", fn _ -> "NULL" end), " )"], else: [] execute = [ "EXPLAIN ", build_explain_opts(opts), "EXECUTE ", @explain_prepared_statement_name, execute_args ] |> IO.iodata_to_binary() deallocate = "DEALLOCATE #{@explain_prepared_statement_name}" {prepare, set, execute, deallocate} end def build_explain_query(query, opts) do ["EXPLAIN ", build_explain_opts(opts), query] |> IO.iodata_to_binary() end defp build_explain_opts([]), do: [] defp build_explain_opts(opts) do {analyze, opts} = Keyword.pop(opts, :analyze) {verbose, opts} = Keyword.pop(opts, :verbose) # Given only ANALYZE or VERBOSE opts we assume the legacy format # to support all Postgres versions, otherwise assume the new # syntax supported since v9.0 case opts do [] -> [ if_do(quote_boolean(analyze) == "TRUE", "ANALYZE "), if_do(quote_boolean(verbose) == "TRUE", "VERBOSE ") ] opts -> opts = ([analyze: analyze, verbose: verbose] ++ opts) |> Enum.reduce([], fn {_, nil}, acc -> acc {:format, value}, acc -> [String.upcase("#{format_to_sql(value)}") | acc] {opt, value}, acc -> [String.upcase("#{opt} #{quote_boolean(value)}") | acc] end) |> Enum.reverse() |> Enum.join(", ") ["( ", opts, " ) "] end end defp fallback_generic_query(conn, queries, opts) do {prepare, set, execute, deallocate} = queries with {:ok, _} <- query(conn, prepare, [], opts), {:ok, _} <- query(conn, set, [], opts), {:ok, result} <- query(conn, execute, [], opts), {:ok, _} <- query(conn, deallocate, [], opts) do {:ok, result} end end ## Query generation binary_ops = [ ==: " = ", !=: " != ", <=: " <= ", >=: " >= ", <: " < ", >: " > ", +: " + ", -: " - ", *: " * ", /: " / ", and: " AND ", or: " OR ", ilike: " ILIKE ", like: " LIKE " ] @binary_ops Keyword.keys(binary_ops) Enum.map(binary_ops, fn {op, str} -> defp handle_call(unquote(op), 2), do: {:binary_op, unquote(str)} end) defp handle_call(fun, _arity), do: {:fun, Atom.to_string(fun)} defp select(%{select: %{fields: fields}} = query, select_distinct, sources) do ["SELECT", select_distinct, ?\s | select_fields(fields, sources, query)] end defp select_fields([], _sources, _query), do: "TRUE" defp select_fields(fields, sources, query) do Enum.map_intersperse(fields, ", ", fn {:&, _, [idx]} -> case elem(sources, idx) do {nil, source, nil} -> error!( query, "PostgreSQL adapter does not support selecting all fields from fragment #{source}. " <> "Please specify exactly which fields you want to select" ) {source, _, nil} -> error!( query, "PostgreSQL adapter does not support selecting all fields from #{source} without a schema. " <> "Please specify a schema or specify exactly which fields you want to select" ) {_, source, _} -> source end {key, value} -> [expr(value, sources, query), " AS " | quote_name(key)] value -> expr(value, sources, query) end) end defp distinct(nil, _, _), do: {[], []} defp distinct(%ByExpr{expr: []}, _, _), do: {[], []} defp distinct(%ByExpr{expr: true}, _, _), do: {" DISTINCT", []} defp distinct(%ByExpr{expr: false}, _, _), do: {[], []} defp distinct(%ByExpr{expr: exprs}, sources, query) do {[ " DISTINCT ON (", Enum.map_intersperse(exprs, ", ", fn {_, expr} -> expr(expr, sources, query) end), ?) ], exprs} end defp from(%{from: %{source: source, hints: hints}} = query, sources) do {from, name} = get_source(query, sources, 0, source) [" FROM ", from, " AS ", name | Enum.map(hints, &[?\s | &1])] end defp cte(%{with_ctes: %WithExpr{queries: [_ | _]}} = query, sources) do %{with_ctes: with} = query recursive_opt = if with.recursive, do: "RECURSIVE ", else: "" ctes = Enum.map_intersperse(with.queries, ", ", &cte_expr(&1, sources, query)) ["WITH ", recursive_opt, ctes, " "] end defp cte(%{with_ctes: _}, _), do: [] defp cte_expr({name, opts, cte}, sources, query) do materialized_opt = case opts[:materialized] do nil -> "" true -> "MATERIALIZED" false -> "NOT MATERIALIZED" end operation_opt = Map.get(opts, :operation) [quote_name(name), " AS ", materialized_opt, cte_query(cte, sources, query, operation_opt)] end defp cte_query(query, sources, parent_query, nil) do cte_query(query, sources, parent_query, :all) end defp cte_query(%Ecto.Query{} = query, sources, parent_query, :update_all) do query = put_in(query.aliases[@parent_as], {parent_query, sources}) ["(", update_all(query), ")"] end defp cte_query(%Ecto.Query{} = query, sources, parent_query, :delete_all) do query = put_in(query.aliases[@parent_as], {parent_query, sources}) ["(", delete_all(query), ")"] end defp cte_query(%Ecto.Query{} = query, _sources, _parent_query, :insert_all) do error!(query, "Postgres adapter does not support CTE operation :insert_all") end defp cte_query(%Ecto.Query{} = query, sources, parent_query, :all) do query = put_in(query.aliases[@parent_as], {parent_query, sources}) ["(", all(query, subquery_as_prefix(sources)), ")"] end defp cte_query(%QueryExpr{expr: expr}, sources, query, _operation) do expr(expr, sources, query) end defp update_fields(%{updates: updates} = query, sources) do for( %{expr: expr} <- updates, {op, kw} <- expr, {key, value} <- kw, do: update_op(op, key, value, sources, query) ) |> Enum.intersperse(", ") end defp update_op(:set, key, value, sources, query) do [quote_name(key), " = " | expr(value, sources, query)] end defp update_op(:inc, key, value, sources, query) do [ quote_name(key), " = ", quote_qualified_name(key, sources, 0), " + " | expr(value, sources, query) ] end defp update_op(:push, key, value, sources, query) do [ quote_name(key), " = array_append(", quote_qualified_name(key, sources, 0), ", ", expr(value, sources, query), ?) ] end defp update_op(:pull, key, value, sources, query) do [ quote_name(key), " = array_remove(", quote_qualified_name(key, sources, 0), ", ", expr(value, sources, query), ?) ] end defp update_op(command, _key, _value, _sources, query) do error!(query, "unknown update operation #{inspect(command)} for PostgreSQL") end defp using_join(%{joins: []}, _kind, _prefix, _sources), do: {[], []} defp using_join(%{joins: joins} = query, :update_all, prefix, sources) do {inner_joins, other_joins} = Enum.split_while(joins, &(&1.qual == :inner)) if inner_joins == [] and other_joins != [] do error!( query, "Need at least one inner join at the beginning to use other joins with update_all" ) end froms = Enum.map_intersperse(inner_joins, ", ", fn %JoinExpr{qual: :inner, ix: ix, source: source} -> {join, name} = get_source(query, sources, ix, source) [join, " AS " | [name]] end) join_clauses = join(%{query | joins: other_joins}, sources) wheres = for %JoinExpr{on: %QueryExpr{expr: value} = expr} <- inner_joins, value != true, do: expr |> Map.put(:__struct__, BooleanExpr) |> Map.put(:op, :and) {[?\s, prefix, ?\s, froms | join_clauses], wheres} end defp using_join(%{joins: joins} = query, kind, prefix, sources) do froms = Enum.map_intersperse(joins, ", ", fn %JoinExpr{qual: :inner, ix: ix, source: source} -> {join, name} = get_source(query, sources, ix, source) [join, " AS " | name] %JoinExpr{qual: qual} -> error!(query, "PostgreSQL supports only inner joins on #{kind}, got: `#{qual}`") end) wheres = for %JoinExpr{on: %QueryExpr{expr: value} = expr} <- joins, value != true, do: expr |> Map.put(:__struct__, BooleanExpr) |> Map.put(:op, :and) {[?\s, prefix, ?\s | froms], wheres} end defp join(%{joins: []}, _sources), do: [] defp join(%{joins: joins} = query, sources) do [ ?\s | Enum.map_intersperse(joins, ?\s, fn %JoinExpr{ on: %QueryExpr{expr: expr}, qual: qual, ix: ix, source: source, hints: hints } -> if hints != [] do error!(query, "table hints are not supported by PostgreSQL") end {join, name} = get_source(query, sources, ix, source) [join_qual(qual, query), join, " AS ", name | join_on(qual, expr, sources, query)] end) ] end defp join_on(:cross, true, _sources, _query), do: [] defp join_on(:cross_lateral, true, _sources, _query), do: [] defp join_on(_qual, expr, sources, query), do: [" ON " | expr(expr, sources, query)] defp join_qual(:inner, _), do: "INNER JOIN " defp join_qual(:inner_lateral, _), do: "INNER JOIN LATERAL " defp join_qual(:left, _), do: "LEFT OUTER JOIN " defp join_qual(:left_lateral, _), do: "LEFT OUTER JOIN LATERAL " defp join_qual(:right, _), do: "RIGHT OUTER JOIN " defp join_qual(:full, _), do: "FULL OUTER JOIN " defp join_qual(:cross, _), do: "CROSS JOIN " defp join_qual(:cross_lateral, _), do: "CROSS JOIN LATERAL " defp join_qual(qual, query), do: error!( query, "join qualifier #{inspect(qual)} is not supported in the PostgreSQL adapter" ) defp where(%{wheres: wheres} = query, sources) do boolean(" WHERE ", wheres, sources, query) end defp having(%{havings: havings} = query, sources) do boolean(" HAVING ", havings, sources, query) end defp group_by(%{group_bys: []}, _sources), do: [] defp group_by(%{group_bys: group_bys} = query, sources) do [ " GROUP BY " | Enum.map_intersperse(group_bys, ", ", fn %ByExpr{expr: expr} -> Enum.map_intersperse(expr, ", ", &expr(&1, sources, query)) end) ] end defp window(%{windows: []}, _sources), do: [] defp window(%{windows: windows} = query, sources) do [ " WINDOW " | Enum.map_intersperse(windows, ", ", fn {name, %{expr: kw}} -> [quote_name(name), " AS " | window_exprs(kw, sources, query)] end) ] end defp window_exprs(kw, sources, query) do [?(, Enum.map_intersperse(kw, ?\s, &window_expr(&1, sources, query)), ?)] end defp window_expr({:partition_by, fields}, sources, query) do ["PARTITION BY " | Enum.map_intersperse(fields, ", ", &expr(&1, sources, query))] end defp window_expr({:order_by, fields}, sources, query) do ["ORDER BY " | Enum.map_intersperse(fields, ", ", &order_by_expr(&1, sources, query))] end defp window_expr({:frame, {:fragment, _, _} = fragment}, sources, query) do expr(fragment, sources, query) end defp order_by(%{order_bys: []}, _distinct, _sources), do: [] defp order_by(%{order_bys: order_bys} = query, distinct, sources) do order_bys = Enum.flat_map(order_bys, & &1.expr) order_bys = order_by_concat(distinct, order_bys) [" ORDER BY " | Enum.map_intersperse(order_bys, ", ", &order_by_expr(&1, sources, query))] end defp order_by_concat([head | left], [head | right]), do: [head | order_by_concat(left, right)] defp order_by_concat(left, right), do: left ++ right defp order_by_expr({dir, expr}, sources, query) do str = expr(expr, sources, query) case dir do :asc -> str :asc_nulls_last -> [str | " ASC NULLS LAST"] :asc_nulls_first -> [str | " ASC NULLS FIRST"] :desc -> [str | " DESC"] :desc_nulls_last -> [str | " DESC NULLS LAST"] :desc_nulls_first -> [str | " DESC NULLS FIRST"] end end defp limit(%{limit: nil}, _sources), do: [] defp limit(%{limit: %{with_ties: true}, order_bys: []} = query, _sources) do error!( query, "PostgreSQL adapter requires an `order_by` clause if the " <> "`:with_ties` limit option is `true`" ) end defp limit(%{limit: %{expr: expr, with_ties: true}} = query, sources) do [" FETCH FIRST ", expr(expr, sources, query) | " ROWS WITH TIES"] end defp limit(%{limit: %{expr: expr}} = query, sources) do [" LIMIT " | expr(expr, sources, query)] end defp offset(%{offset: nil}, _sources), do: [] defp offset(%{offset: %QueryExpr{expr: expr}} = query, sources) do [" OFFSET " | expr(expr, sources, query)] end defp combinations(%{combinations: combinations}, as_prefix) do Enum.map(combinations, fn {:union, query} -> [" UNION (", all(query, as_prefix), ")"] {:union_all, query} -> [" UNION ALL (", all(query, as_prefix), ")"] {:except, query} -> [" EXCEPT (", all(query, as_prefix), ")"] {:except_all, query} -> [" EXCEPT ALL (", all(query, as_prefix), ")"] {:intersect, query} -> [" INTERSECT (", all(query, as_prefix), ")"] {:intersect_all, query} -> [" INTERSECT ALL (", all(query, as_prefix), ")"] end) end defp lock(%{lock: nil}, _sources), do: [] defp lock(%{lock: binary}, _sources) when is_binary(binary), do: [?\s | binary] defp lock(%{lock: expr} = query, sources), do: [?\s | expr(expr, sources, query)] defp boolean(_name, [], _sources, _query), do: [] defp boolean(name, [%{expr: expr, op: op} | query_exprs], sources, query) do [ name | Enum.reduce(query_exprs, {op, paren_expr(expr, sources, query)}, fn %BooleanExpr{expr: expr, op: op}, {op, acc} -> {op, [acc, operator_to_boolean(op), paren_expr(expr, sources, query)]} %BooleanExpr{expr: expr, op: op}, {_, acc} -> {op, [?(, acc, ?), operator_to_boolean(op), paren_expr(expr, sources, query)]} end) |> elem(1) ] end defp operator_to_boolean(:and), do: " AND " defp operator_to_boolean(:or), do: " OR " defp parens_for_select([first_expr | _] = expr) do if is_binary(first_expr) and String.match?(first_expr, ~r/^\s*select\s/i) do [?(, expr, ?)] else expr end end defp paren_expr(expr, sources, query) do [?(, expr(expr, sources, query), ?)] end defp expr({:^, [], [ix]}, _sources, _query) do [?$ | Integer.to_string(ix + 1)] end defp expr({{:., _, [{:parent_as, _, [as]}, field]}, _, []}, _sources, query) when is_atom(field) do {ix, sources} = get_parent_sources_ix(query, as) quote_qualified_name(field, sources, ix) end defp expr({{:., _, [{:&, _, [idx]}, field]}, _, []}, sources, _query) when is_atom(field) do quote_qualified_name(field, sources, idx) end defp expr({:&, _, [idx]}, sources, _query) do {_, source, _} = elem(sources, idx) source end defp expr({:in, _, [_left, []]}, _sources, _query) do "false" end defp expr({:in, _, [left, right]}, sources, query) when is_list(right) do args = Enum.map_intersperse(right, ?,, &expr(&1, sources, query)) [expr(left, sources, query), " IN (", args, ?)] end defp expr({:in, _, [left, {:^, _, [ix, _]}]}, sources, query) do [expr(left, sources, query), " = ANY($", Integer.to_string(ix + 1), ?)] end defp expr({:in, _, [left, %Ecto.SubQuery{} = subquery]}, sources, query) do [expr(left, sources, query), " IN ", expr(subquery, sources, query)] end defp expr({:in, _, [left, right]}, sources, query) do [expr(left, sources, query), " = ANY(", expr(right, sources, query), ?)] end defp expr({:is_nil, _, [arg]}, sources, query) do [expr(arg, sources, query) | " IS NULL"] end defp expr({:not, _, [expr]}, sources, query) do ["NOT (", expr(expr, sources, query), ?)] end defp expr(%Ecto.SubQuery{query: query}, sources, parent_query) do combinations = Enum.map(query.combinations, fn {type, combination_query} -> {type, put_in(combination_query.aliases[@parent_as], {parent_query, sources})} end) query = put_in(query.combinations, combinations) query = put_in(query.aliases[@parent_as], {parent_query, sources}) [?(, all(query, subquery_as_prefix(sources)), ?)] end defp expr({:fragment, _, [kw]}, _sources, query) when is_list(kw) or tuple_size(kw) == 3 do error!(query, "PostgreSQL adapter does not support keyword or interpolated fragments") end defp expr({:fragment, _, parts}, sources, query) do Enum.map(parts, fn {:raw, part} -> part {:expr, expr} -> expr(expr, sources, query) end) |> parens_for_select end defp expr({:values, _, [types, idx, num_rows]}, _, _query) do [?(, values_list(types, idx + 1, num_rows), ?)] end defp expr({:literal, _, [literal]}, _sources, _query) do quote_name(literal) end defp expr({:splice, _, [{:^, _, [idx, length]}]}, _sources, _query) do Enum.map_join(1..length, ",", &"$#{idx + &1}") end defp expr({:selected_as, _, [name]}, _sources, _query) do [quote_name(name)] end defp expr({:datetime_add, _, [datetime, count, interval]}, sources, query) do [ expr(datetime, sources, query), type_unless_typed(datetime, "timestamp"), " + ", interval(count, interval, sources, query) ] end defp expr({:date_add, _, [date, count, interval]}, sources, query) do [ ?(, expr(date, sources, query), type_unless_typed(date, "date"), " + ", interval(count, interval, sources, query) | ")::date" ] end defp expr({:json_extract_path, _, [expr, path]}, sources, query) do json_extract_path(expr, path, sources, query) end defp expr({:filter, _, [agg, filter]}, sources, query) do aggregate = expr(agg, sources, query) [aggregate, " FILTER (WHERE ", expr(filter, sources, query), ?)] end defp expr({:over, _, [agg, name]}, sources, query) when is_atom(name) do aggregate = expr(agg, sources, query) [aggregate, " OVER " | quote_name(name)] end defp expr({:over, _, [agg, kw]}, sources, query) do aggregate = expr(agg, sources, query) [aggregate, " OVER ", window_exprs(kw, sources, query)] end defp expr({:{}, _, elems}, sources, query) do [?(, Enum.map_intersperse(elems, ?,, &expr(&1, sources, query)), ?)] end defp expr({:count, _, []}, _sources, _query), do: "count(*)" defp expr({:==, _, [{:json_extract_path, _, [expr, path]} = left, right]}, sources, query) when is_binary(right) or is_integer(right) or is_boolean(right) do case Enum.split(path, -1) do {path, [last]} when is_binary(last) -> extracted = json_extract_path(expr, path, sources, query) [?(, extracted, "@>'{", escape_json(last), ": ", escape_json(right) | "}')"] _ -> [maybe_paren(left, sources, query), " = " | maybe_paren(right, sources, query)] end end defp expr({fun, _, args}, sources, query) when is_atom(fun) and is_list(args) do {modifier, args} = case args do [rest, :distinct] -> {"DISTINCT ", [rest]} _ -> {[], args} end case handle_call(fun, length(args)) do {:binary_op, op} -> [left, right] = args [maybe_paren(left, sources, query), op | maybe_paren(right, sources, query)] {:fun, fun} -> [fun, ?(, modifier, Enum.map_intersperse(args, ", ", &expr(&1, sources, query)), ?)] end end defp expr([], _sources, _query) do # We cannot compare in postgres with the empty array # i. e. `where array_column = ARRAY[];` # as that will result in an error: # ERROR: cannot determine type of empty array # HINT: Explicitly cast to the desired type, for example ARRAY[]::integer[]. # # On the other side comparing with '{}' works # because '{}' represents the pseudo-type "unknown" # and thus the type gets inferred based on the column # it is being compared to so `where array_column = '{}';` works. "'{}'" end defp expr(list, sources, query) when is_list(list) do ["ARRAY[", Enum.map_intersperse(list, ?,, &expr(&1, sources, query)), ?]] end defp expr(%Decimal{} = decimal, _sources, _query) do Decimal.to_string(decimal, :normal) end defp expr(%Ecto.Query.Tagged{value: binary, type: :binary}, _sources, _query) when is_binary(binary) do ["'\\x", Base.encode16(binary, case: :lower) | "'::bytea"] end defp expr(%Ecto.Query.Tagged{value: bitstring, type: :bitstring}, _sources, _query) when is_bitstring(bitstring) do bitstring_literal(bitstring) end defp expr(%Ecto.Query.Tagged{value: other, type: type}, sources, query) do [maybe_paren(other, sources, query), ?:, ?: | tagged_to_db(type)] end defp expr(nil, _sources, _query), do: "NULL" defp expr(true, _sources, _query), do: "TRUE" defp expr(false, _sources, _query), do: "FALSE" defp expr(literal, _sources, _query) when is_binary(literal) do [?\', escape_string(literal), ?\'] end defp expr(literal, _sources, _query) when is_integer(literal) do Integer.to_string(literal) end defp expr(literal, _sources, _query) when is_float(literal) do [Float.to_string(literal) | "::float"] end defp expr(expr, _sources, query) do error!(query, "unsupported expression: #{inspect(expr)}") end defp json_extract_path(expr, [], sources, query) do expr(expr, sources, query) end defp json_extract_path(expr, path, sources, query) do path = Enum.map_intersperse(path, ?,, &escape_json/1) [?(, expr(expr, sources, query), "#>'{", path, "}')"] end defp values_list(types, idx, num_rows) do rows = Enum.to_list(1..num_rows) [ "VALUES ", intersperse_reduce(rows, ?,, idx, fn _, idx -> {value, idx} = values_expr(types, idx) {[?(, value, ?)], idx} end) |> elem(0) ] end defp values_expr(types, idx) do intersperse_reduce(types, ?,, idx, fn {_field, type}, idx -> {[?$, Integer.to_string(idx), ?:, ?: | tagged_to_db(type)], idx + 1} end) end defp type_unless_typed(%Ecto.Query.Tagged{}, _type), do: [] defp type_unless_typed(_, type), do: [?:, ?: | type] # Always use the largest possible type for integers defp tagged_to_db(:id), do: "bigint" defp tagged_to_db(:integer), do: "bigint" defp tagged_to_db({:array, type}), do: [tagged_to_db(type), ?[, ?]] defp tagged_to_db(type), do: ecto_to_db(type) defp interval(count, interval, _sources, _query) when is_integer(count) do ["interval '", String.Chars.Integer.to_string(count), ?\s, interval, ?\'] end defp interval(count, interval, _sources, _query) when is_float(count) do count = :erlang.float_to_binary(count, [:compact, decimals: 16]) ["interval '", count, ?\s, interval, ?\'] end defp interval(count, interval, sources, query) do [?(, expr(count, sources, query), "::numeric * ", interval(1, interval, sources, query), ?)] end defp maybe_paren({op, _, [_, _]} = expr, sources, query) when op in @binary_ops, do: paren_expr(expr, sources, query) defp maybe_paren({:is_nil, _, [_]} = expr, sources, query), do: paren_expr(expr, sources, query) defp maybe_paren(expr, sources, query), do: expr(expr, sources, query) defp returning(%{select: nil}, _sources), do: [] defp returning(%{select: %{fields: fields}} = query, sources), do: [" RETURNING " | select_fields(fields, sources, query)] defp returning([]), do: [] defp returning(returning), do: [" RETURNING " | quote_names(returning)] defp create_names(%{sources: sources}, as_prefix) do create_names(sources, 0, tuple_size(sources), as_prefix) |> List.to_tuple() end defp create_names(sources, pos, limit, as_prefix) when pos < limit do [create_name(sources, pos, as_prefix) | create_names(sources, pos + 1, limit, as_prefix)] end defp create_names(_sources, pos, pos, as_prefix) do [as_prefix] end defp subquery_as_prefix(sources) do [?s | :erlang.element(tuple_size(sources), sources)] end defp create_name(sources, pos, as_prefix) do case elem(sources, pos) do {:fragment, _, _} -> {nil, as_prefix ++ [?f | Integer.to_string(pos)], nil} {:values, _, _} -> {nil, as_prefix ++ [?v | Integer.to_string(pos)], nil} {table, schema, prefix} -> name = as_prefix ++ [create_alias(table) | Integer.to_string(pos)] {quote_name(prefix, table), name, schema} %Ecto.SubQuery{} -> {nil, as_prefix ++ [?s | Integer.to_string(pos)], nil} end end defp create_alias(<>) when first in ?a..?z when first in ?A..?Z do first end defp create_alias(_) do ?t end # DDL alias Ecto.Migration.{Table, Index, Reference, Constraint} @creates [:create, :create_if_not_exists] @drops [:drop, :drop_if_exists] @impl true def execute_ddl({command, %Table{} = table, columns}) when command in @creates do table_name = quote_name(table.prefix, table.name) query = [ "CREATE TABLE ", if_do(command == :create_if_not_exists, "IF NOT EXISTS "), table_name, ?\s, ?(, column_definitions(table, columns), pk_definition(columns, ", "), ?), options_expr(table.options) ] [query] ++ comments_on("TABLE", table_name, table.comment) ++ comments_for_columns(table_name, columns) end def execute_ddl({command, %Table{} = table, mode}) when command in @drops do [ [ "DROP TABLE ", if_do(command == :drop_if_exists, "IF EXISTS "), quote_name(table.prefix, table.name), drop_mode(mode) ] ] end def execute_ddl({:alter, %Table{} = table, changes}) do table_name = quote_name(table.prefix, table.name) query = [ "ALTER TABLE ", table_name, ?\s, column_changes(table, changes), pk_definition(changes, ", ADD ") ] [query] ++ comments_on("TABLE", table_name, table.comment) ++ comments_for_columns(table_name, changes) end def execute_ddl({command, %Index{} = index}) when command in @creates do fields = Enum.map_intersperse(index.columns, ", ", &index_expr/1) include_fields = Enum.map_intersperse(index.include, ", ", &index_expr/1) maybe_nulls_distinct = case index.nulls_distinct do nil -> [] true -> " NULLS DISTINCT" false -> " NULLS NOT DISTINCT" end queries = [ [ "CREATE ", if_do(index.unique, "UNIQUE "), "INDEX ", if_do(index.concurrently, "CONCURRENTLY "), if_do(command == :create_if_not_exists, "IF NOT EXISTS "), quote_name(index.name), " ON ", if_do(index.only, "ONLY "), quote_name(index.prefix, index.table), if_do(index.using, [" USING ", to_string(index.using)]), ?\s, ?(, fields, ?), if_do(include_fields != [], [" INCLUDE ", ?(, include_fields, ?)]), maybe_nulls_distinct, if_do(index.options != nil, [" WITH ", ?(, index.options, ?)]), if_do(index.where, [" WHERE ", to_string(index.where)]) ] ] queries ++ comments_on("INDEX", quote_name(index.prefix, index.name), index.comment) end def execute_ddl({command, %Index{} = index, mode}) when command in @drops do [ [ "DROP INDEX ", if_do(index.concurrently, "CONCURRENTLY "), if_do(command == :drop_if_exists, "IF EXISTS "), quote_name(index.prefix, index.name), drop_mode(mode) ] ] end def execute_ddl({:rename, %Index{} = current_index, new_name}) do [ [ "ALTER INDEX ", quote_name(current_index.prefix, current_index.name), " RENAME TO ", quote_name(new_name) ] ] end def execute_ddl({:rename, %Table{} = current_table, %Table{} = new_table}) do [ [ "ALTER TABLE ", quote_name(current_table.prefix, current_table.name), " RENAME TO ", quote_name(nil, new_table.name) ] ] end def execute_ddl({:rename, %Table{} = table, current_column, new_column}) do [ [ "ALTER TABLE ", quote_name(table.prefix, table.name), " RENAME ", quote_name(current_column), " TO ", quote_name(new_column) ] ] end def execute_ddl({:create, %Constraint{} = constraint}) do table_name = quote_name(constraint.prefix, constraint.table) queries = [["ALTER TABLE ", table_name, " ADD ", new_constraint_expr(constraint)]] queries ++ comments_on("CONSTRAINT", constraint.name, constraint.comment, table_name) end def execute_ddl({command, %Constraint{} = constraint, mode}) when command in @drops do [ [ "ALTER TABLE ", quote_name(constraint.prefix, constraint.table), " DROP CONSTRAINT ", if_do(command == :drop_if_exists, "IF EXISTS "), quote_name(constraint.name), drop_mode(mode) ] ] end def execute_ddl(string) when is_binary(string), do: [string] def execute_ddl(keyword) when is_list(keyword), do: error!(nil, "PostgreSQL adapter does not support keyword lists in execute") @impl true def ddl_logs(%Postgrex.Result{} = result) do %{messages: messages} = result for message <- messages do %{message: message, severity: severity} = message {ddl_log_level(severity), message, []} end end @impl true def table_exists_query(table) do {"SELECT true FROM information_schema.tables WHERE table_name = $1 AND table_schema = current_schema() LIMIT 1", [table]} end defp drop_mode(:cascade), do: " CASCADE" defp drop_mode(:restrict), do: [] # From https://www.postgresql.org/docs/current/protocol-error-fields.html. defp ddl_log_level("DEBUG"), do: :debug defp ddl_log_level("LOG"), do: :info defp ddl_log_level("INFO"), do: :info defp ddl_log_level("NOTICE"), do: :info defp ddl_log_level("WARNING"), do: :warn defp ddl_log_level("ERROR"), do: :error defp ddl_log_level("FATAL"), do: :error defp ddl_log_level("PANIC"), do: :error defp ddl_log_level(_severity), do: :info defp pk_definition(columns, prefix) do pks = for {action, name, _, opts} <- columns, action != :remove, opts[:primary_key], do: name case pks do [] -> [] _ -> [prefix, "PRIMARY KEY (", quote_names(pks), ")"] end end defp comments_on(_object, _name, nil), do: [] defp comments_on(object, name, comment) do [["COMMENT ON ", object, ?\s, name, " IS ", single_quote(comment)]] end defp comments_on(_object, _name, nil, _table_name), do: [] defp comments_on(object, name, comment, table_name) do [ [ "COMMENT ON ", object, ?\s, quote_name(name), " ON ", table_name, " IS ", single_quote(comment) ] ] end defp comments_for_columns(table_name, columns) do Enum.flat_map(columns, fn {:remove, _column_name, _column_type, _opts} -> [] {_operation, column_name, _column_type, opts} -> column_name = [table_name, ?. | quote_name(column_name)] comments_on("COLUMN", column_name, opts[:comment]) _ -> [] end) end defp column_definitions(table, columns) do Enum.map_intersperse(columns, ", ", &column_definition(table, &1)) end defp column_definition(table, {:add, name, %Reference{} = ref, opts}) do [ quote_name(name), ?\s, reference_column_type(ref.type, opts), column_options(ref.type, opts), ", ", reference_expr(ref, table, name) ] end defp column_definition(_table, {:add, name, type, opts}) do [quote_name(name), ?\s, column_type(type, opts), column_options(type, opts)] end defp column_changes(table, columns) do Enum.map_intersperse(columns, ", ", &column_change(table, &1)) end defp column_change(table, {:add, name, %Reference{} = ref, opts}) do [ "ADD COLUMN ", quote_name(name), ?\s, reference_column_type(ref.type, opts), column_options(ref.type, opts), ", ADD ", reference_expr(ref, table, name) ] end defp column_change(_table, {:add, name, type, opts}) do ["ADD COLUMN ", quote_name(name), ?\s, column_type(type, opts), column_options(type, opts)] end defp column_change(table, {:add_if_not_exists, name, %Reference{} = ref, opts}) do [ "ADD COLUMN IF NOT EXISTS ", quote_name(name), ?\s, reference_column_type(ref.type, opts), column_options(ref.type, opts), ", ADD ", reference_expr(ref, table, name) ] end defp column_change(_table, {:add_if_not_exists, name, type, opts}) do [ "ADD COLUMN IF NOT EXISTS ", quote_name(name), ?\s, column_type(type, opts), column_options(type, opts) ] end defp column_change(table, {:modify, name, %Reference{} = ref, opts}) do [ drop_reference_expr(opts[:from], table, name), "ALTER COLUMN ", quote_name(name), " TYPE ", reference_column_type(ref.type, opts), ", ADD ", reference_expr(ref, table, name), modify_null(name, opts), modify_default(name, ref.type, opts) ] end defp column_change(table, {:modify, name, type, opts}) do [ drop_reference_expr(opts[:from], table, name), "ALTER COLUMN ", quote_name(name), " TYPE ", column_type(type, opts), modify_null(name, opts), modify_default(name, type, opts) ] end defp column_change(_table, {:remove, name}), do: ["DROP COLUMN ", quote_name(name)] defp column_change(table, {:remove, name, %Reference{} = ref, _opts}) do [drop_reference_expr(ref, table, name), "DROP COLUMN ", quote_name(name)] end defp column_change(_table, {:remove, name, _type, _opts}), do: ["DROP COLUMN ", quote_name(name)] defp column_change(table, {:remove_if_exists, name, %Reference{} = ref}) do [ drop_reference_if_exists_expr(ref, table, name), "DROP COLUMN IF EXISTS ", quote_name(name) ] end defp column_change(table, {:remove_if_exists, name, _type}), do: column_change(table, {:remove_if_exists, name}) defp column_change(_table, {:remove_if_exists, name}), do: ["DROP COLUMN IF EXISTS ", quote_name(name)] defp modify_null(name, opts) do case Keyword.get(opts, :null) do true -> [", ALTER COLUMN ", quote_name(name), " DROP NOT NULL"] false -> [", ALTER COLUMN ", quote_name(name), " SET NOT NULL"] nil -> [] end end defp modify_default(name, type, opts) do case Keyword.fetch(opts, :default) do {:ok, val} -> [", ALTER COLUMN ", quote_name(name), " SET", default_expr({:ok, val}, type)] :error -> [] end end defp column_options(type, opts) do default = Keyword.fetch(opts, :default) null = Keyword.get(opts, :null) [default_expr(default, type), null_expr(null)] end defp null_expr(false), do: " NOT NULL" defp null_expr(true), do: " NULL" defp null_expr(_), do: [] defp new_constraint_expr(%Constraint{check: check} = constraint) when is_binary(check) do [ "CONSTRAINT ", quote_name(constraint.name), " CHECK (", check, ")", validate(constraint.validate) ] end defp new_constraint_expr(%Constraint{exclude: exclude} = constraint) when is_binary(exclude) do [ "CONSTRAINT ", quote_name(constraint.name), " EXCLUDE USING ", exclude, validate(constraint.validate) ] end defp default_expr({:ok, nil}, _type), do: " DEFAULT NULL" defp default_expr({:ok, literal}, type), do: [" DEFAULT ", default_type(literal, type)] defp default_expr(:error, _), do: [] defp default_type(list, {:array, inner} = type) when is_list(list) do [ "ARRAY[", Enum.map_intersperse(list, ?,, &default_type(&1, inner)), "]::", ecto_to_db(type) ] end defp default_type(literal, _type) when is_binary(literal) do if :binary.match(literal, <<0>>) == :nomatch and String.valid?(literal) do single_quote(literal) else encoded = "\\x" <> Base.encode16(literal, case: :lower) raise ArgumentError, "default values are interpolated as UTF-8 strings and cannot contain null bytes. " <> "`#{inspect(literal)}` is invalid. If you want to write it as a binary, use \"#{encoded}\", " <> "otherwise refer to PostgreSQL documentation for instructions on how to escape this SQL type" end end defp default_type(literal, _type) when is_bitstring(literal) do bitstring_literal(literal) end defp default_type(literal, _type) when is_number(literal), do: to_string(literal) defp default_type(literal, _type) when is_boolean(literal), do: to_string(literal) defp default_type(%{} = map, :map) do library = Application.get_env(:postgrex, :json_library, Jason) default = IO.iodata_to_binary(library.encode_to_iodata!(map)) [single_quote(default)] end defp default_type({:fragment, expr}, _type), do: [expr] defp default_type(expr, type), do: raise( ArgumentError, "unknown default `#{inspect(expr)}` for type `#{inspect(type)}`. " <> ":default may be a string, number, boolean, list of strings, list of integers, map (when type is Map), or a fragment(...)" ) defp index_expr(literal) when is_binary(literal), do: literal defp index_expr(literal), do: quote_name(literal) defp options_expr(nil), do: [] defp options_expr(keyword) when is_list(keyword), do: error!(nil, "PostgreSQL adapter does not support keyword lists in :options") defp options_expr(options), do: [?\s, options] defp column_type({:array, type}, opts), do: [column_type(type, opts), "[]"] defp column_type(type, opts) when type in ~w(time utc_datetime naive_datetime)a do generated = Keyword.get(opts, :generated) [ecto_to_db(type), "(0)", generated_expr(generated)] end defp column_type(type, opts) when type in ~w(time_usec utc_datetime_usec naive_datetime_usec)a do precision = Keyword.get(opts, :precision) generated = Keyword.get(opts, :generated) type_name = ecto_to_db(type) type = if precision do [type_name, ?(, to_string(precision), ?)] else type_name end [type, generated_expr(generated)] end defp column_type(:identity, opts) do start_value = [Keyword.get(opts, :start_value)] increment = [Keyword.get(opts, :increment)] generated = Keyword.get(opts, :generated) type_name = ecto_to_db(:identity) if generated do [type_name, generated_expr(generated)] else cleanup = fn v -> is_integer(v) and v > 0 end sequence = start_value |> Enum.filter(cleanup) |> Enum.map(&"START WITH #{&1}") |> Kernel.++( increment |> Enum.filter(cleanup) |> Enum.map(&"INCREMENT BY #{&1}") ) case sequence do [] -> [type_name, " GENERATED BY DEFAULT AS IDENTITY"] _ -> [type_name, " GENERATED BY DEFAULT AS IDENTITY(", Enum.join(sequence, " "), ") "] end end end defp column_type(:duration, opts) do precision = Keyword.get(opts, :precision) fields = Keyword.get(opts, :fields) generated = Keyword.get(opts, :generated) type_name = ecto_to_db(:duration) type = cond do fields && precision -> [type_name, " ", fields, ?(, to_string(precision), ?)] precision -> [type_name, ?(, to_string(precision), ?)] fields -> [type_name, " ", fields] true -> [type_name] end [type, generated_expr(generated)] end defp column_type(type, opts) do size = Keyword.get(opts, :size) precision = Keyword.get(opts, :precision) scale = Keyword.get(opts, :scale) generated = Keyword.get(opts, :generated) type_name = ecto_to_db(type) type = cond do size -> [type_name, ?(, to_string(size), ?)] precision -> [type_name, ?(, to_string(precision), ?,, to_string(scale || 0), ?)] type == :string -> [type_name, "(255)"] true -> type_name end [type, generated_expr(generated)] end defp generated_expr(nil), do: [] defp generated_expr(expr) when is_binary(expr) do [" GENERATED ", expr] end defp generated_expr(other) do raise ArgumentError, "the `:generated` option only accepts strings, received: #{inspect(other)}" end defp reference_expr(%Reference{} = ref, table, name) do {current_columns, reference_columns} = Enum.unzip([{name, ref.column} | ref.with]) [ "CONSTRAINT ", reference_name(ref, table, name), ?\s, "FOREIGN KEY (", quote_names(current_columns), ") REFERENCES ", quote_name(Keyword.get(ref.options, :prefix, table.prefix), ref.table), ?(, quote_names(reference_columns), ?), reference_match(ref.match), reference_on_delete(ref.on_delete), reference_on_update(ref.on_update), validate(ref.validate) ] end defp drop_reference_expr({%Reference{} = ref, _opts}, table, name), do: drop_reference_expr(ref, table, name) defp drop_reference_expr(%Reference{} = ref, table, name), do: ["DROP CONSTRAINT ", reference_name(ref, table, name), ", "] defp drop_reference_expr(_, _, _), do: [] defp drop_reference_if_exists_expr(%Reference{} = ref, table, name), do: ["DROP CONSTRAINT IF EXISTS ", reference_name(ref, table, name), ", "] defp drop_reference_if_exists_expr(_, _, _), do: [] defp reference_name(%Reference{name: nil}, table, column), do: quote_name("#{table.name}_#{column}_fkey") defp reference_name(%Reference{name: name}, _table, _column), do: quote_name(name) defp reference_column_type(:serial, _opts), do: "integer" defp reference_column_type(:bigserial, _opts), do: "bigint" defp reference_column_type(:identity, _opts), do: "bigint" defp reference_column_type(type, opts), do: column_type(type, opts) defp reference_on_delete(:nilify_all), do: " ON DELETE SET NULL" defp reference_on_delete({:nilify, columns}), do: [" ON DELETE SET NULL (", quote_names(columns), ")"] defp reference_on_delete(:delete_all), do: " ON DELETE CASCADE" defp reference_on_delete(:restrict), do: " ON DELETE RESTRICT" defp reference_on_delete(_), do: [] defp reference_on_update(:nilify_all), do: " ON UPDATE SET NULL" defp reference_on_update(:update_all), do: " ON UPDATE CASCADE" defp reference_on_update(:restrict), do: " ON UPDATE RESTRICT" defp reference_on_update(_), do: [] defp reference_match(nil), do: [] defp reference_match(:full), do: " MATCH FULL" defp reference_match(:simple), do: " MATCH SIMPLE" defp reference_match(:partial), do: " MATCH PARTIAL" defp validate(false), do: " NOT VALID" defp validate(_), do: [] ## Helpers defp get_source(query, sources, ix, source) do {expr, name, _schema} = elem(sources, ix) name = maybe_add_column_names(source, name) {expr || expr(source, sources, query), name} end defp get_parent_sources_ix(query, as) do case query.aliases[@parent_as] do {%{aliases: %{^as => ix}}, sources} -> {ix, sources} {%{} = parent, _sources} -> get_parent_sources_ix(parent, as) end end defp maybe_add_column_names({:values, _, [types, _, _]}, name) do fields = Keyword.keys(types) [name, ?\s, ?(, quote_names(fields), ?)] end defp maybe_add_column_names(_, name), do: name defp quote_qualified_name(name, sources, ix) do {_, source, _} = elem(sources, ix) [source, ?. | quote_name(name)] end defp quote_names(names) do Enum.map_intersperse(names, ?,, "e_name/1) end defp quote_name(nil, name), do: quote_name(name) defp quote_name(prefix, name), do: [quote_name(prefix), ?., quote_name(name)] defp quote_name(name) when is_atom(name) do quote_name(Atom.to_string(name)) end defp quote_name(name) when is_binary(name) do if String.contains?(name, "\"") do error!(nil, "bad literal/field/index/table name #{inspect(name)} (\" is not permitted)") end [?", name, ?"] end # TRUE, ON, or 1 to enable the option, and FALSE, OFF, or 0 to disable it defp quote_boolean(nil), do: nil defp quote_boolean(true), do: "TRUE" defp quote_boolean(false), do: "FALSE" defp quote_boolean(value), do: error!(nil, "bad boolean value #{value}") defp format_to_sql(:text), do: "FORMAT TEXT" defp format_to_sql(:map), do: "FORMAT JSON" defp format_to_sql(:yaml), do: "FORMAT YAML" defp single_quote(value), do: [?', escape_string(value), ?'] defp bitstring_literal(value) do size = bit_size(value) <> = value [?b, ?', val |> Integer.to_string(2) |> String.pad_leading(size, ["0"]), ?'] end defp intersperse_reduce(list, separator, user_acc, reducer, acc \\ []) defp intersperse_reduce([], _separator, user_acc, _reducer, acc), do: {acc, user_acc} defp intersperse_reduce([elem], _separator, user_acc, reducer, acc) do {elem, user_acc} = reducer.(elem, user_acc) {[acc | elem], user_acc} end defp intersperse_reduce([elem | rest], separator, user_acc, reducer, acc) do {elem, user_acc} = reducer.(elem, user_acc) intersperse_reduce(rest, separator, user_acc, reducer, [acc, elem, separator]) end defp if_do(condition, value) do if condition, do: value, else: [] end defp escape_string(value) when is_binary(value) do :binary.replace(value, "'", "''", [:global]) end defp escape_json(value) when is_binary(value) do escaped = value |> escape_string() |> :binary.replace("\"", "\\\"", [:global]) [?", escaped, ?"] end defp escape_json(value) when is_integer(value) do Integer.to_string(value) end defp escape_json(true), do: ["true"] defp escape_json(false), do: ["false"] defp ecto_to_db({:array, t}), do: [ecto_to_db(t), ?[, ?]] defp ecto_to_db(:id), do: "integer" defp ecto_to_db(:identity), do: "bigint" defp ecto_to_db(:serial), do: "serial" defp ecto_to_db(:bigserial), do: "bigserial" defp ecto_to_db(:binary_id), do: "uuid" defp ecto_to_db(:string), do: "varchar" defp ecto_to_db(:bitstring), do: "varbit" defp ecto_to_db(:binary), do: "bytea" defp ecto_to_db(:map), do: Application.fetch_env!(:ecto_sql, :postgres_map_type) defp ecto_to_db({:map, _}), do: Application.fetch_env!(:ecto_sql, :postgres_map_type) defp ecto_to_db(:time_usec), do: "time" defp ecto_to_db(:utc_datetime), do: "timestamp" defp ecto_to_db(:utc_datetime_usec), do: "timestamp" defp ecto_to_db(:naive_datetime), do: "timestamp" defp ecto_to_db(:naive_datetime_usec), do: "timestamp" defp ecto_to_db(:duration), do: "interval" defp ecto_to_db(atom) when is_atom(atom), do: Atom.to_string(atom) defp ecto_to_db(type) do raise ArgumentError, "unsupported type `#{inspect(type)}`. The type can either be an atom, a string " <> "or a tuple of the form `{:map, t}` or `{:array, t}` where `t` itself follows the same conditions." end defp error!(nil, message) do raise ArgumentError, message end defp error!(query, message) do raise Ecto.QueryError, query: query, message: message end end end