* feat(api): usability testing (#1686) * feat(api): usability testing - wip * feat(db): usabiity testing * feat(api): usability testing - api * feat(api): usability testing - api * feat(api): usability testing - db change * feat(api): usability testing - db change * feat(api): usability testing - unit tests update * feat(api): usability testing - test and tasks stats * feat(api): usability testing - sessions list fix, return zeros if test id is not having signals * Api v1.16.0 (#1698) * feat: canvas support [assist] (#1641) * feat(tracker/ui): start canvas support * feat(tracker): slpeer -> peerjs for canvas streams * fix(ui): fix agent canvas peer id * fix(ui): fix agent canvas peer id * fix(ui): fix peer removal * feat(tracker): canvas recorder * feat(tracker): canvas recorder * feat(tracker): canvas recorder * feat(tracker): canvas recorder * feat(ui): canvas support for ui * fix(tracker): fix falling tests * feat(ui): replay canvas in video * feat(ui): refactor video streaming to draw on canvas * feat(ui): 10hz check for canvas replay * feat(ui): fix for tests * feat(ui): fix for tests * feat(ui): fix for tests * feat(ui): fix for tests cov * feat(ui): mroe test coverage * fix(ui): styling * fix(tracker): support backend settings for canvas * feat(ui): allow devtools to be resizeable (#1605) * fix(ui): console redux tab null check * Api v1.15.0 (#1689) * fix(chalice): fix create alert with MS Teams notification channel closes openreplay/openreplay#1677 * fix(chalice): fix MS Teams notifications * refactor(chalice): enhanced MS Teams notifications closes openreplay/openreplay#1681 (cherry picked from commit265897f509) * fix(ui): filter keys conflcit with metadata, path analysis 4 col * fix(ui): clear the filers and series on card type change * fix(player): fix msg reader bug * fix(DB): fix CH wrong version (#1692) (cherry picked from commit48dbbb55db) * fix(ui): filter keys conflcit with metadata * fix(tracker): unique broadcast channel name * fix(chalice): fixed delete cards (#1697) (cherry picked from commit92fedd310c) * fix(tracker): add trycatch to ignore iframe errors * feat(backend): added ARM arch support to backend services [Dockerfile] * feat(backend): removed userAgent from sessions and unstarted-sessions tables * fix(DB): change path-analysis card size --------- Co-authored-by: Delirium <nikita@openreplay.com> Co-authored-by: Shekar Siri <sshekarsiri@gmail.com> Co-authored-by: Alexander <zavorotynskiy@pm.me> * refactor(chalice): cleaned code (#1699) * feat(api): usability testing - added start_path to the resposne, remove count from the list * feat(api): usability testing - test to have response count and live count * feat(api): usability testing - test to have additional data * Revert "refactor(chalice): cleaned code (#1699)" (#1702) This reverts commit83f2b0c12c. * feat(api): usability testing - responses with total and other improvements * change(api): vulnerability whitelist udpate * feat(api): usability testing - create added missing columns, and sessions with user_id search * feat(api): usability testing - update test with responseCount * feat(api): usability testing - timestamps in unix * feat(api): usability testing - request with proper case change * feat(api): usability testing - task.description nullable * feat(api): usability testing - check deleted status * Api v1.16.0 (#1707) * fix(chalice): fixed search sessions * fix(chalice): fixed search sessions * refactor(chalice): upgraded dependencies * refactor(crons): upgraded dependencies * refactor(alerts): upgraded dependencies * Api v1.16.0 (#1712) * feat(DB): user-testing support * feat(chalice): user testing support * feat(chalice): support utxVideo (#1726) * feat(chalice): changed bucket name for ux testing webcamera videos --------- Co-authored-by: Shekar Siri <sshekarsiri@gmail.com> Co-authored-by: Kraiem Taha Yassine <tahayk2@gmail.com> Co-authored-by: Delirium <nikita@openreplay.com> Co-authored-by: Alexander <zavorotynskiy@pm.me>
1289 lines
73 KiB
Python
1289 lines
73 KiB
Python
import logging
|
|
from typing import List
|
|
|
|
import schemas
|
|
from chalicelib.core import events, metadata, projects, performance_event, sessions_favorite
|
|
from chalicelib.utils import pg_client, helper, metrics_helper
|
|
from chalicelib.utils import sql_helper as sh
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
SESSION_PROJECTION_BASE_COLS = """s.project_id,
|
|
s.session_id::text AS session_id,
|
|
s.user_uuid,
|
|
s.user_id,
|
|
s.user_os,
|
|
s.user_browser,
|
|
s.user_device,
|
|
s.user_device_type,
|
|
s.user_country,
|
|
s.user_city,
|
|
s.user_state,
|
|
s.start_ts,
|
|
s.duration,
|
|
s.events_count,
|
|
s.pages_count,
|
|
s.errors_count,
|
|
s.user_anonymous_id,
|
|
s.platform,
|
|
s.issue_score,
|
|
s.timezone,
|
|
to_jsonb(s.issue_types) AS issue_types """
|
|
|
|
SESSION_PROJECTION_COLS = SESSION_PROJECTION_BASE_COLS + """,
|
|
favorite_sessions.session_id NOTNULL AS favorite,
|
|
COALESCE((SELECT TRUE
|
|
FROM public.user_viewed_sessions AS fs
|
|
WHERE s.session_id = fs.session_id
|
|
AND fs.user_id = %(userId)s LIMIT 1), FALSE) AS viewed """
|
|
|
|
|
|
# This function executes the query and return result
|
|
def search_sessions(data: schemas.SessionsSearchPayloadSchema, project_id, user_id, errors_only=False,
|
|
error_status=schemas.ErrorStatus.all, count_only=False, issue=None, ids_only=False,
|
|
platform="web"):
|
|
if data.bookmarked:
|
|
data.startTimestamp, data.endTimestamp = sessions_favorite.get_start_end_timestamp(project_id, user_id)
|
|
|
|
full_args, query_part = search_query_parts(data=data, error_status=error_status, errors_only=errors_only,
|
|
favorite_only=data.bookmarked, issue=issue, project_id=project_id,
|
|
user_id=user_id, platform=platform)
|
|
if data.limit is not None and data.page is not None:
|
|
full_args["sessions_limit"] = data.limit
|
|
full_args["sessions_limit_s"] = (data.page - 1) * data.limit
|
|
full_args["sessions_limit_e"] = data.page * data.limit
|
|
else:
|
|
full_args["sessions_limit"] = 200
|
|
full_args["sessions_limit_s"] = 0
|
|
full_args["sessions_limit_e"] = 200
|
|
|
|
meta_keys = []
|
|
with pg_client.PostgresClient() as cur:
|
|
if errors_only:
|
|
main_query = cur.mogrify(f"""SELECT DISTINCT er.error_id,
|
|
COALESCE((SELECT TRUE
|
|
FROM public.user_viewed_errors AS ve
|
|
WHERE er.error_id = ve.error_id
|
|
AND ve.user_id = %(userId)s LIMIT 1), FALSE) AS viewed
|
|
{query_part};""", full_args)
|
|
|
|
elif count_only:
|
|
main_query = cur.mogrify(f"""SELECT COUNT(DISTINCT s.session_id) AS count_sessions,
|
|
COUNT(DISTINCT s.user_uuid) AS count_users
|
|
{query_part};""", full_args)
|
|
elif data.group_by_user:
|
|
g_sort = "count(full_sessions)"
|
|
if data.order is None:
|
|
data.order = schemas.SortOrderType.desc.value
|
|
else:
|
|
data.order = data.order
|
|
if data.sort is not None and data.sort != 'sessionsCount':
|
|
sort = helper.key_to_snake_case(data.sort)
|
|
g_sort = f"{'MIN' if data.order == schemas.SortOrderType.desc else 'MAX'}({sort})"
|
|
else:
|
|
sort = 'start_ts'
|
|
|
|
meta_keys = metadata.get(project_id=project_id)
|
|
main_query = cur.mogrify(f"""SELECT COUNT(*) AS count,
|
|
COALESCE(JSONB_AGG(users_sessions)
|
|
FILTER (WHERE rn>%(sessions_limit_s)s AND rn<=%(sessions_limit_e)s), '[]'::JSONB) AS sessions
|
|
FROM (SELECT user_id,
|
|
count(full_sessions) AS user_sessions_count,
|
|
jsonb_agg(full_sessions) FILTER (WHERE rn <= 1) AS last_session,
|
|
MIN(full_sessions.start_ts) AS first_session_ts,
|
|
ROW_NUMBER() OVER (ORDER BY {g_sort} {data.order}) AS rn
|
|
FROM (SELECT *, ROW_NUMBER() OVER (PARTITION BY user_id ORDER BY {sort} {data.order}) AS rn
|
|
FROM (SELECT DISTINCT ON(s.session_id) {SESSION_PROJECTION_COLS}
|
|
{"," if len(meta_keys) > 0 else ""}{",".join([f'metadata_{m["index"]}' for m in meta_keys])}
|
|
{query_part}
|
|
) AS filtred_sessions
|
|
) AS full_sessions
|
|
GROUP BY user_id
|
|
) AS users_sessions;""",
|
|
full_args)
|
|
elif ids_only:
|
|
main_query = cur.mogrify(f"""SELECT DISTINCT ON(s.session_id) s.session_id
|
|
{query_part}
|
|
ORDER BY s.session_id desc
|
|
LIMIT %(sessions_limit)s OFFSET %(sessions_limit_s)s;""",
|
|
full_args)
|
|
else:
|
|
if data.order is None:
|
|
data.order = schemas.SortOrderType.desc.value
|
|
else:
|
|
data.order = data.order
|
|
sort = 'session_id'
|
|
if data.sort is not None and data.sort != "session_id":
|
|
# sort += " " + data.order + "," + helper.key_to_snake_case(data.sort)
|
|
sort = helper.key_to_snake_case(data.sort)
|
|
|
|
meta_keys = metadata.get(project_id=project_id)
|
|
main_query = cur.mogrify(f"""SELECT COUNT(full_sessions) AS count,
|
|
COALESCE(JSONB_AGG(full_sessions)
|
|
FILTER (WHERE rn>%(sessions_limit_s)s AND rn<=%(sessions_limit_e)s), '[]'::JSONB) AS sessions
|
|
FROM (SELECT *, ROW_NUMBER() OVER (ORDER BY {sort} {data.order}, issue_score DESC) AS rn
|
|
FROM (SELECT DISTINCT ON(s.session_id) {SESSION_PROJECTION_COLS}
|
|
{"," if len(meta_keys) > 0 else ""}{",".join([f'metadata_{m["index"]}' for m in meta_keys])}
|
|
{query_part}
|
|
ORDER BY s.session_id desc) AS filtred_sessions
|
|
ORDER BY {sort} {data.order}, issue_score DESC) AS full_sessions;""",
|
|
full_args)
|
|
logging.debug("--------------------")
|
|
logging.debug(main_query)
|
|
logging.debug("--------------------")
|
|
try:
|
|
cur.execute(main_query)
|
|
sessions = cur.fetchone()
|
|
except Exception as err:
|
|
logging.warning("--------- SESSIONS SEARCH QUERY EXCEPTION -----------")
|
|
logging.warning(main_query.decode('UTF-8'))
|
|
logging.warning("--------- PAYLOAD -----------")
|
|
logging.warning(data.model_dump_json())
|
|
logging.warning("--------------------")
|
|
raise err
|
|
if errors_only or ids_only:
|
|
return helper.list_to_camel_case(cur.fetchall())
|
|
|
|
if count_only:
|
|
return helper.dict_to_camel_case(sessions)
|
|
|
|
total = sessions["count"]
|
|
sessions = sessions["sessions"]
|
|
|
|
if data.group_by_user:
|
|
for i, s in enumerate(sessions):
|
|
sessions[i] = {**s.pop("last_session")[0], **s}
|
|
sessions[i].pop("rn")
|
|
sessions[i]["metadata"] = {k["key"]: sessions[i][f'metadata_{k["index"]}'] for k in meta_keys \
|
|
if sessions[i][f'metadata_{k["index"]}'] is not None}
|
|
else:
|
|
for i, s in enumerate(sessions):
|
|
sessions[i]["metadata"] = {k["key"]: sessions[i][f'metadata_{k["index"]}'] for k in meta_keys \
|
|
if sessions[i][f'metadata_{k["index"]}'] is not None}
|
|
# if not data.group_by_user and data.sort is not None and data.sort != "session_id":
|
|
# sessions = sorted(sessions, key=lambda s: s[helper.key_to_snake_case(data.sort)],
|
|
# reverse=data.order.upper() == "DESC")
|
|
return {
|
|
'total': total,
|
|
'sessions': helper.list_to_camel_case(sessions)
|
|
}
|
|
|
|
|
|
# TODO: remove "table of" search from this function
|
|
def search2_series(data: schemas.SessionsSearchPayloadSchema, project_id: int, density: int,
|
|
view_type: schemas.MetricTimeseriesViewType, metric_type: schemas.MetricType,
|
|
metric_of: schemas.MetricOfTable, metric_value: List):
|
|
step_size = int(metrics_helper.__get_step_size(endTimestamp=data.endTimestamp, startTimestamp=data.startTimestamp,
|
|
density=density, factor=1, decimal=True))
|
|
extra_event = None
|
|
if metric_of == schemas.MetricOfTable.visited_url:
|
|
extra_event = "events.pages"
|
|
elif metric_of == schemas.MetricOfTable.issues and len(metric_value) > 0:
|
|
data.filters.append(schemas.SessionSearchFilterSchema(value=metric_value, type=schemas.FilterType.issue,
|
|
operator=schemas.SearchEventOperator._is))
|
|
full_args, query_part = search_query_parts(data=data, error_status=None, errors_only=False,
|
|
favorite_only=False, issue=None, project_id=project_id,
|
|
user_id=None, extra_event=extra_event)
|
|
full_args["step_size"] = step_size
|
|
sessions = []
|
|
with pg_client.PostgresClient() as cur:
|
|
if metric_type == schemas.MetricType.timeseries:
|
|
if view_type == schemas.MetricTimeseriesViewType.line_chart:
|
|
main_query = cur.mogrify(f"""WITH full_sessions AS (SELECT DISTINCT ON(s.session_id) s.session_id, s.start_ts
|
|
{query_part})
|
|
SELECT generated_timestamp AS timestamp,
|
|
COUNT(s) AS count
|
|
FROM generate_series(%(startDate)s, %(endDate)s, %(step_size)s) AS generated_timestamp
|
|
LEFT JOIN LATERAL ( SELECT 1 AS s
|
|
FROM full_sessions
|
|
WHERE start_ts >= generated_timestamp
|
|
AND start_ts <= generated_timestamp + %(step_size)s) AS sessions ON (TRUE)
|
|
GROUP BY generated_timestamp
|
|
ORDER BY generated_timestamp;""", full_args)
|
|
else:
|
|
main_query = cur.mogrify(f"""SELECT count(DISTINCT s.session_id) AS count
|
|
{query_part};""", full_args)
|
|
|
|
logging.debug("--------------------")
|
|
logging.debug(main_query)
|
|
logging.debug("--------------------")
|
|
try:
|
|
cur.execute(main_query)
|
|
except Exception as err:
|
|
logging.warning("--------- SESSIONS-SERIES QUERY EXCEPTION -----------")
|
|
logging.warning(main_query.decode('UTF-8'))
|
|
logging.warning("--------- PAYLOAD -----------")
|
|
logging.warning(data.model_dump_json())
|
|
logging.warning("--------------------")
|
|
raise err
|
|
if view_type == schemas.MetricTimeseriesViewType.line_chart:
|
|
sessions = cur.fetchall()
|
|
else:
|
|
sessions = cur.fetchone()["count"]
|
|
elif metric_type == schemas.MetricType.table:
|
|
if isinstance(metric_of, schemas.MetricOfTable):
|
|
main_col = "user_id"
|
|
extra_col = ""
|
|
extra_where = ""
|
|
pre_query = ""
|
|
distinct_on = "s.session_id"
|
|
if metric_of == schemas.MetricOfTable.user_country:
|
|
main_col = "user_country"
|
|
elif metric_of == schemas.MetricOfTable.user_device:
|
|
main_col = "user_device"
|
|
elif metric_of == schemas.MetricOfTable.user_browser:
|
|
main_col = "user_browser"
|
|
elif metric_of == schemas.MetricOfTable.issues:
|
|
main_col = "issue"
|
|
extra_col = f", UNNEST(s.issue_types) AS {main_col}"
|
|
if len(metric_value) > 0:
|
|
extra_where = []
|
|
for i in range(len(metric_value)):
|
|
arg_name = f"selected_issue_{i}"
|
|
extra_where.append(f"{main_col} = %({arg_name})s")
|
|
full_args[arg_name] = metric_value[i]
|
|
extra_where = f"WHERE ({' OR '.join(extra_where)})"
|
|
elif metric_of == schemas.MetricOfTable.visited_url:
|
|
main_col = "path"
|
|
extra_col = ", path"
|
|
distinct_on += ",path"
|
|
main_query = cur.mogrify(f"""{pre_query}
|
|
SELECT COUNT(*) AS count,
|
|
SUM(users_sessions.session_count) AS total_sessions,
|
|
COALESCE(JSONB_AGG(users_sessions) FILTER ( WHERE rn <= 200 ), '[]'::JSONB) AS values
|
|
FROM (SELECT {main_col} AS name,
|
|
count(DISTINCT session_id) AS session_count,
|
|
ROW_NUMBER() OVER (ORDER BY count(full_sessions) DESC) AS rn
|
|
FROM (SELECT *
|
|
FROM (SELECT DISTINCT ON({distinct_on}) s.session_id, s.user_uuid,
|
|
s.user_id, s.user_os,
|
|
s.user_browser, s.user_device,
|
|
s.user_device_type, s.user_country, s.issue_types{extra_col}
|
|
{query_part}
|
|
ORDER BY s.session_id desc) AS filtred_sessions
|
|
) AS full_sessions
|
|
{extra_where}
|
|
GROUP BY {main_col}
|
|
ORDER BY session_count DESC) AS users_sessions;""",
|
|
full_args)
|
|
logging.debug("--------------------")
|
|
logging.debug(main_query)
|
|
logging.debug("--------------------")
|
|
cur.execute(main_query)
|
|
sessions = helper.dict_to_camel_case(cur.fetchone())
|
|
for s in sessions["values"]:
|
|
s.pop("rn")
|
|
|
|
return sessions
|
|
|
|
|
|
def search2_table(data: schemas.SessionsSearchPayloadSchema, project_id: int, density: int,
|
|
metric_of: schemas.MetricOfTable, metric_value: List):
|
|
step_size = int(metrics_helper.__get_step_size(endTimestamp=data.endTimestamp, startTimestamp=data.startTimestamp,
|
|
density=density, factor=1, decimal=True))
|
|
extra_event = None
|
|
if metric_of == schemas.MetricOfTable.visited_url:
|
|
extra_event = "events.pages"
|
|
elif metric_of == schemas.MetricOfTable.issues and len(metric_value) > 0:
|
|
data.filters.append(schemas.SessionSearchFilterSchema(value=metric_value, type=schemas.FilterType.issue,
|
|
operator=schemas.SearchEventOperator._is))
|
|
full_args, query_part = search_query_parts(data=data, error_status=None, errors_only=False,
|
|
favorite_only=False, issue=None, project_id=project_id,
|
|
user_id=None, extra_event=extra_event)
|
|
full_args["step_size"] = step_size
|
|
with pg_client.PostgresClient() as cur:
|
|
if isinstance(metric_of, schemas.MetricOfTable):
|
|
main_col = "user_id"
|
|
extra_col = ""
|
|
extra_where = ""
|
|
pre_query = ""
|
|
distinct_on = "s.session_id"
|
|
if metric_of == schemas.MetricOfTable.user_country:
|
|
main_col = "user_country"
|
|
elif metric_of == schemas.MetricOfTable.user_device:
|
|
main_col = "user_device"
|
|
elif metric_of == schemas.MetricOfTable.user_browser:
|
|
main_col = "user_browser"
|
|
elif metric_of == schemas.MetricOfTable.issues:
|
|
main_col = "issue"
|
|
extra_col = f", UNNEST(s.issue_types) AS {main_col}"
|
|
if len(metric_value) > 0:
|
|
extra_where = []
|
|
for i in range(len(metric_value)):
|
|
arg_name = f"selected_issue_{i}"
|
|
extra_where.append(f"{main_col} = %({arg_name})s")
|
|
full_args[arg_name] = metric_value[i]
|
|
extra_where = f"WHERE ({' OR '.join(extra_where)})"
|
|
elif metric_of == schemas.MetricOfTable.visited_url:
|
|
main_col = "path"
|
|
extra_col = ", path"
|
|
distinct_on += ",path"
|
|
main_query = cur.mogrify(f"""{pre_query}
|
|
SELECT COUNT(*) AS count,
|
|
COALESCE(SUM(users_sessions.session_count),0) AS total_sessions,
|
|
COALESCE(JSONB_AGG(users_sessions) FILTER ( WHERE rn <= 200 ), '[]'::JSONB) AS values
|
|
FROM (SELECT {main_col} AS name,
|
|
count(DISTINCT session_id) AS session_count,
|
|
ROW_NUMBER() OVER (ORDER BY count(full_sessions) DESC) AS rn
|
|
FROM (SELECT *
|
|
FROM (SELECT DISTINCT ON({distinct_on}) s.session_id, s.user_uuid,
|
|
s.user_id, s.user_os,
|
|
s.user_browser, s.user_device,
|
|
s.user_device_type, s.user_country, s.issue_types{extra_col}
|
|
{query_part}
|
|
ORDER BY s.session_id desc) AS filtred_sessions
|
|
) AS full_sessions
|
|
{extra_where}
|
|
GROUP BY {main_col}
|
|
ORDER BY session_count DESC) AS users_sessions;""",
|
|
full_args)
|
|
logging.debug("--------------------")
|
|
logging.debug(main_query)
|
|
logging.debug("--------------------")
|
|
cur.execute(main_query)
|
|
sessions = helper.dict_to_camel_case(cur.fetchone())
|
|
for s in sessions["values"]:
|
|
s.pop("rn")
|
|
|
|
return sessions
|
|
|
|
|
|
def search_table_of_individual_issues(data: schemas.SessionsSearchPayloadSchema, project_id: int):
|
|
full_args, query_part = search_query_parts(data=data, error_status=None, errors_only=False,
|
|
favorite_only=False, issue=None, project_id=project_id,
|
|
user_id=None)
|
|
|
|
with pg_client.PostgresClient() as cur:
|
|
full_args["issues_limit"] = data.limit
|
|
full_args["issues_limit_s"] = (data.page - 1) * data.limit
|
|
full_args["issues_limit_e"] = data.page * data.limit
|
|
main_query = cur.mogrify(f"""SELECT COUNT(1) AS count,
|
|
COALESCE(SUM(session_count), 0) AS total_sessions,
|
|
COALESCE(JSONB_AGG(ranked_issues)
|
|
FILTER ( WHERE rn > %(issues_limit_s)s
|
|
AND rn <= %(issues_limit_e)s ), '[]'::JSONB) AS values
|
|
FROM (SELECT *, ROW_NUMBER() OVER (ORDER BY session_count DESC) AS rn
|
|
FROM (SELECT type AS name, context_string AS value, COUNT(DISTINCT session_id) AS session_count
|
|
FROM (SELECT session_id
|
|
{query_part}) AS filtered_sessions
|
|
INNER JOIN events_common.issues USING (session_id)
|
|
INNER JOIN public.issues USING (issue_id)
|
|
WHERE project_id = %(project_id)s
|
|
AND timestamp >= %(startDate)s
|
|
AND timestamp <= %(endDate)s
|
|
GROUP BY type, context_string
|
|
ORDER BY session_count DESC) AS filtered_issues
|
|
) AS ranked_issues;""", full_args)
|
|
logging.debug("--------------------")
|
|
logging.debug(main_query)
|
|
logging.debug("--------------------")
|
|
cur.execute(main_query)
|
|
sessions = helper.dict_to_camel_case(cur.fetchone())
|
|
for s in sessions["values"]:
|
|
s.pop("rn")
|
|
|
|
return sessions
|
|
|
|
|
|
def __is_valid_event(is_any: bool, event: schemas.SessionSearchEventSchema2):
|
|
return not (not is_any and len(event.value) == 0 and event.type not in [schemas.EventType.request_details,
|
|
schemas.EventType.graphql] \
|
|
or event.type in [schemas.PerformanceEventType.location_dom_complete,
|
|
schemas.PerformanceEventType.location_largest_contentful_paint_time,
|
|
schemas.PerformanceEventType.location_ttfb,
|
|
schemas.PerformanceEventType.location_avg_cpu_load,
|
|
schemas.PerformanceEventType.location_avg_memory_usage
|
|
] and (event.source is None or len(event.source) == 0) \
|
|
or event.type in [schemas.EventType.request_details, schemas.EventType.graphql] and (
|
|
event.filters is None or len(event.filters) == 0))
|
|
|
|
|
|
# this function generates the query and return the generated-query with the dict of query arguments
|
|
def search_query_parts(data: schemas.SessionsSearchPayloadSchema, error_status, errors_only, favorite_only, issue,
|
|
project_id, user_id, platform="web", extra_event=None):
|
|
ss_constraints = []
|
|
full_args = {"project_id": project_id, "startDate": data.startTimestamp, "endDate": data.endTimestamp,
|
|
"projectId": project_id, "userId": user_id}
|
|
extra_constraints = [
|
|
"s.project_id = %(project_id)s",
|
|
"s.duration IS NOT NULL"
|
|
]
|
|
extra_from = ""
|
|
events_query_part = ""
|
|
issues = []
|
|
if len(data.filters) > 0:
|
|
meta_keys = None
|
|
for i, f in enumerate(data.filters):
|
|
if not isinstance(f.value, list):
|
|
f.value = [f.value]
|
|
filter_type = f.type
|
|
f.value = helper.values_for_operator(value=f.value, op=f.operator)
|
|
f_k = f"f_value{i}"
|
|
full_args = {**full_args, **sh.multi_values(f.value, value_key=f_k)}
|
|
op = sh.get_sql_operator(f.operator) \
|
|
if filter_type not in [schemas.FilterType.events_count] else f.operator
|
|
is_any = sh.isAny_opreator(f.operator)
|
|
is_undefined = sh.isUndefined_operator(f.operator)
|
|
if not is_any and not is_undefined and len(f.value) == 0:
|
|
continue
|
|
is_not = False
|
|
if sh.is_negation_operator(f.operator):
|
|
is_not = True
|
|
if filter_type == schemas.FilterType.user_browser:
|
|
if is_any:
|
|
extra_constraints.append('s.user_browser IS NOT NULL')
|
|
ss_constraints.append('ms.user_browser IS NOT NULL')
|
|
else:
|
|
extra_constraints.append(
|
|
sh.multi_conditions(f's.user_browser {op} %({f_k})s', f.value, is_not=is_not, value_key=f_k))
|
|
ss_constraints.append(
|
|
sh.multi_conditions(f'ms.user_browser {op} %({f_k})s', f.value, is_not=is_not,
|
|
value_key=f_k))
|
|
|
|
elif filter_type in [schemas.FilterType.user_os, schemas.FilterType.user_os_ios]:
|
|
if is_any:
|
|
extra_constraints.append('s.user_os IS NOT NULL')
|
|
ss_constraints.append('ms.user_os IS NOT NULL')
|
|
else:
|
|
extra_constraints.append(
|
|
sh.multi_conditions(f's.user_os {op} %({f_k})s', f.value, is_not=is_not, value_key=f_k))
|
|
ss_constraints.append(
|
|
sh.multi_conditions(f'ms.user_os {op} %({f_k})s', f.value, is_not=is_not, value_key=f_k))
|
|
|
|
elif filter_type in [schemas.FilterType.user_device, schemas.FilterType.user_device_ios]:
|
|
if is_any:
|
|
extra_constraints.append('s.user_device IS NOT NULL')
|
|
ss_constraints.append('ms.user_device IS NOT NULL')
|
|
else:
|
|
extra_constraints.append(
|
|
sh.multi_conditions(f's.user_device {op} %({f_k})s', f.value, is_not=is_not, value_key=f_k))
|
|
ss_constraints.append(
|
|
sh.multi_conditions(f'ms.user_device {op} %({f_k})s', f.value, is_not=is_not, value_key=f_k))
|
|
|
|
elif filter_type in [schemas.FilterType.user_country, schemas.FilterType.user_country_ios]:
|
|
if is_any:
|
|
extra_constraints.append('s.user_country IS NOT NULL')
|
|
ss_constraints.append('ms.user_country IS NOT NULL')
|
|
else:
|
|
extra_constraints.append(
|
|
sh.multi_conditions(f's.user_country {op} %({f_k})s', f.value, is_not=is_not, value_key=f_k))
|
|
ss_constraints.append(
|
|
sh.multi_conditions(f'ms.user_country {op} %({f_k})s', f.value, is_not=is_not,
|
|
value_key=f_k))
|
|
|
|
elif filter_type == schemas.FilterType.user_city:
|
|
if is_any:
|
|
extra_constraints.append('s.user_city IS NOT NULL')
|
|
ss_constraints.append('ms.user_city IS NOT NULL')
|
|
else:
|
|
extra_constraints.append(
|
|
sh.multi_conditions(f's.user_city {op} %({f_k})s', f.value, is_not=is_not, value_key=f_k))
|
|
ss_constraints.append(
|
|
sh.multi_conditions(f'ms.user_city {op} %({f_k})s', f.value, is_not=is_not,
|
|
value_key=f_k))
|
|
|
|
elif filter_type == schemas.FilterType.user_state:
|
|
if is_any:
|
|
extra_constraints.append('s.user_state IS NOT NULL')
|
|
ss_constraints.append('ms.user_state IS NOT NULL')
|
|
else:
|
|
extra_constraints.append(
|
|
sh.multi_conditions(f's.user_state {op} %({f_k})s', f.value, is_not=is_not, value_key=f_k))
|
|
ss_constraints.append(
|
|
sh.multi_conditions(f'ms.user_state {op} %({f_k})s', f.value, is_not=is_not,
|
|
value_key=f_k))
|
|
|
|
elif filter_type in [schemas.FilterType.utm_source]:
|
|
if is_any:
|
|
extra_constraints.append('s.utm_source IS NOT NULL')
|
|
ss_constraints.append('ms.utm_source IS NOT NULL')
|
|
elif is_undefined:
|
|
extra_constraints.append('s.utm_source IS NULL')
|
|
ss_constraints.append('ms.utm_source IS NULL')
|
|
else:
|
|
extra_constraints.append(
|
|
sh.multi_conditions(f's.utm_source {op} %({f_k})s::text', f.value, is_not=is_not,
|
|
value_key=f_k))
|
|
ss_constraints.append(
|
|
sh.multi_conditions(f'ms.utm_source {op} %({f_k})s::text', f.value, is_not=is_not,
|
|
value_key=f_k))
|
|
elif filter_type in [schemas.FilterType.utm_medium]:
|
|
if is_any:
|
|
extra_constraints.append('s.utm_medium IS NOT NULL')
|
|
ss_constraints.append('ms.utm_medium IS NOT NULL')
|
|
elif is_undefined:
|
|
extra_constraints.append('s.utm_medium IS NULL')
|
|
ss_constraints.append('ms.utm_medium IS NULL')
|
|
else:
|
|
extra_constraints.append(
|
|
sh.multi_conditions(f's.utm_medium {op} %({f_k})s::text', f.value, is_not=is_not,
|
|
value_key=f_k))
|
|
ss_constraints.append(
|
|
sh.multi_conditions(f'ms.utm_medium {op} %({f_k})s::text', f.value, is_not=is_not,
|
|
value_key=f_k))
|
|
elif filter_type in [schemas.FilterType.utm_campaign]:
|
|
if is_any:
|
|
extra_constraints.append('s.utm_campaign IS NOT NULL')
|
|
ss_constraints.append('ms.utm_campaign IS NOT NULL')
|
|
elif is_undefined:
|
|
extra_constraints.append('s.utm_campaign IS NULL')
|
|
ss_constraints.append('ms.utm_campaign IS NULL')
|
|
else:
|
|
extra_constraints.append(
|
|
sh.multi_conditions(f's.utm_campaign {op} %({f_k})s::text', f.value, is_not=is_not,
|
|
value_key=f_k))
|
|
ss_constraints.append(
|
|
sh.multi_conditions(f'ms.utm_campaign {op} %({f_k})s::text', f.value, is_not=is_not,
|
|
value_key=f_k))
|
|
|
|
elif filter_type == schemas.FilterType.duration:
|
|
if len(f.value) > 0 and f.value[0] is not None:
|
|
extra_constraints.append("s.duration >= %(minDuration)s")
|
|
ss_constraints.append("ms.duration >= %(minDuration)s")
|
|
full_args["minDuration"] = f.value[0]
|
|
if len(f.value) > 1 and f.value[1] is not None and int(f.value[1]) > 0:
|
|
extra_constraints.append("s.duration <= %(maxDuration)s")
|
|
ss_constraints.append("ms.duration <= %(maxDuration)s")
|
|
full_args["maxDuration"] = f.value[1]
|
|
elif filter_type == schemas.FilterType.referrer:
|
|
# extra_from += f"INNER JOIN {events.event_type.LOCATION.table} AS p USING(session_id)"
|
|
if is_any:
|
|
extra_constraints.append('s.base_referrer IS NOT NULL')
|
|
else:
|
|
extra_constraints.append(
|
|
sh.multi_conditions(f"s.base_referrer {op} %({f_k})s", f.value, is_not=is_not,
|
|
value_key=f_k))
|
|
elif filter_type == events.EventType.METADATA.ui_type:
|
|
# get metadata list only if you need it
|
|
if meta_keys is None:
|
|
meta_keys = metadata.get(project_id=project_id)
|
|
meta_keys = {m["key"]: m["index"] for m in meta_keys}
|
|
if f.source in meta_keys.keys():
|
|
if is_any:
|
|
extra_constraints.append(f"s.{metadata.index_to_colname(meta_keys[f.source])} IS NOT NULL")
|
|
ss_constraints.append(f"ms.{metadata.index_to_colname(meta_keys[f.source])} IS NOT NULL")
|
|
elif is_undefined:
|
|
extra_constraints.append(f"s.{metadata.index_to_colname(meta_keys[f.source])} IS NULL")
|
|
ss_constraints.append(f"ms.{metadata.index_to_colname(meta_keys[f.source])} IS NULL")
|
|
else:
|
|
extra_constraints.append(
|
|
sh.multi_conditions(
|
|
f"s.{metadata.index_to_colname(meta_keys[f.source])} {op} %({f_k})s::text",
|
|
f.value, is_not=is_not, value_key=f_k))
|
|
ss_constraints.append(
|
|
sh.multi_conditions(
|
|
f"ms.{metadata.index_to_colname(meta_keys[f.source])} {op} %({f_k})s::text",
|
|
f.value, is_not=is_not, value_key=f_k))
|
|
elif filter_type in [schemas.FilterType.user_id, schemas.FilterType.user_id_ios]:
|
|
if is_any:
|
|
extra_constraints.append('s.user_id IS NOT NULL')
|
|
ss_constraints.append('ms.user_id IS NOT NULL')
|
|
elif is_undefined:
|
|
extra_constraints.append('s.user_id IS NULL')
|
|
ss_constraints.append('ms.user_id IS NULL')
|
|
else:
|
|
extra_constraints.append(
|
|
sh.multi_conditions(f"s.user_id {op} %({f_k})s::text", f.value, is_not=is_not,
|
|
value_key=f_k))
|
|
ss_constraints.append(
|
|
sh.multi_conditions(f"ms.user_id {op} %({f_k})s::text", f.value, is_not=is_not,
|
|
value_key=f_k))
|
|
elif filter_type in [schemas.FilterType.user_anonymous_id,
|
|
schemas.FilterType.user_anonymous_id_ios]:
|
|
if is_any:
|
|
extra_constraints.append('s.user_anonymous_id IS NOT NULL')
|
|
ss_constraints.append('ms.user_anonymous_id IS NOT NULL')
|
|
elif is_undefined:
|
|
extra_constraints.append('s.user_anonymous_id IS NULL')
|
|
ss_constraints.append('ms.user_anonymous_id IS NULL')
|
|
else:
|
|
extra_constraints.append(
|
|
sh.multi_conditions(f"s.user_anonymous_id {op} %({f_k})s::text", f.value, is_not=is_not,
|
|
value_key=f_k))
|
|
ss_constraints.append(
|
|
sh.multi_conditions(f"ms.user_anonymous_id {op} %({f_k})s::text", f.value, is_not=is_not,
|
|
value_key=f_k))
|
|
elif filter_type in [schemas.FilterType.rev_id, schemas.FilterType.rev_id_ios]:
|
|
if is_any:
|
|
extra_constraints.append('s.rev_id IS NOT NULL')
|
|
ss_constraints.append('ms.rev_id IS NOT NULL')
|
|
elif is_undefined:
|
|
extra_constraints.append('s.rev_id IS NULL')
|
|
ss_constraints.append('ms.rev_id IS NULL')
|
|
else:
|
|
extra_constraints.append(
|
|
sh.multi_conditions(f"s.rev_id {op} %({f_k})s::text", f.value, is_not=is_not, value_key=f_k))
|
|
ss_constraints.append(
|
|
sh.multi_conditions(f"ms.rev_id {op} %({f_k})s::text", f.value, is_not=is_not,
|
|
value_key=f_k))
|
|
elif filter_type == schemas.FilterType.platform:
|
|
# op = __ sh.get_sql_operator(f.operator)
|
|
extra_constraints.append(
|
|
sh.multi_conditions(f"s.user_device_type {op} %({f_k})s", f.value, is_not=is_not,
|
|
value_key=f_k))
|
|
ss_constraints.append(
|
|
sh.multi_conditions(f"ms.user_device_type {op} %({f_k})s", f.value, is_not=is_not,
|
|
value_key=f_k))
|
|
elif filter_type == schemas.FilterType.issue:
|
|
if is_any:
|
|
extra_constraints.append("array_length(s.issue_types, 1) > 0")
|
|
ss_constraints.append("array_length(ms.issue_types, 1) > 0")
|
|
else:
|
|
if f.source:
|
|
issues.append(f)
|
|
|
|
extra_constraints.append(
|
|
sh.multi_conditions(f"%({f_k})s {op} ANY (s.issue_types)", f.value, is_not=is_not,
|
|
value_key=f_k))
|
|
ss_constraints.append(
|
|
sh.multi_conditions(f"%({f_k})s {op} ANY (ms.issue_types)", f.value, is_not=is_not,
|
|
value_key=f_k))
|
|
|
|
elif filter_type == schemas.FilterType.events_count:
|
|
extra_constraints.append(
|
|
sh.multi_conditions(f"s.events_count {op} %({f_k})s", f.value, is_not=is_not,
|
|
value_key=f_k))
|
|
ss_constraints.append(
|
|
sh.multi_conditions(f"ms.events_count {op} %({f_k})s", f.value, is_not=is_not,
|
|
value_key=f_k))
|
|
# ---------------------------------------------------------------------------
|
|
if len(data.events) > 0:
|
|
valid_events_count = 0
|
|
for event in data.events:
|
|
is_any = sh.isAny_opreator(event.operator)
|
|
if not isinstance(event.value, list):
|
|
event.value = [event.value]
|
|
if __is_valid_event(is_any=is_any, event=event):
|
|
valid_events_count += 1
|
|
events_query_from = []
|
|
event_index = 0
|
|
or_events = data.events_order == schemas.SearchEventOrder._or
|
|
# events_joiner = " FULL JOIN " if or_events else " INNER JOIN LATERAL "
|
|
events_joiner = " UNION " if or_events else " INNER JOIN LATERAL "
|
|
for i, event in enumerate(data.events):
|
|
event_type = event.type
|
|
is_any = sh.isAny_opreator(event.operator)
|
|
if not isinstance(event.value, list):
|
|
event.value = [event.value]
|
|
if not __is_valid_event(is_any=is_any, event=event):
|
|
continue
|
|
op = sh.get_sql_operator(event.operator)
|
|
is_not = False
|
|
if sh.is_negation_operator(event.operator):
|
|
is_not = True
|
|
op = sh.reverse_sql_operator(op)
|
|
if event_index == 0 or or_events:
|
|
event_from = "%s INNER JOIN public.sessions AS ms USING (session_id)"
|
|
event_where = ["ms.project_id = %(projectId)s", "main.timestamp >= %(startDate)s",
|
|
"main.timestamp <= %(endDate)s", "ms.start_ts >= %(startDate)s",
|
|
"ms.start_ts <= %(endDate)s", "ms.duration IS NOT NULL"]
|
|
if favorite_only and not errors_only:
|
|
event_from += "INNER JOIN public.user_favorite_sessions AS fs USING(session_id)"
|
|
event_where.append("fs.user_id = %(userId)s")
|
|
else:
|
|
event_from = "%s"
|
|
event_where = ["main.timestamp >= %(startDate)s", "main.timestamp <= %(endDate)s",
|
|
"main.session_id=event_0.session_id"]
|
|
if data.events_order == schemas.SearchEventOrder._then:
|
|
event_where.append(f"event_{event_index - 1}.timestamp <= main.timestamp")
|
|
e_k = f"e_value{i}"
|
|
s_k = e_k + "_source"
|
|
|
|
event.value = helper.values_for_operator(value=event.value, op=event.operator)
|
|
full_args = {**full_args,
|
|
**sh.multi_values(event.value, value_key=e_k),
|
|
**sh.multi_values(event.source, value_key=s_k)}
|
|
|
|
if event_type == events.EventType.CLICK.ui_type:
|
|
if platform == "web":
|
|
event_from = event_from % f"{events.EventType.CLICK.table} AS main "
|
|
if not is_any:
|
|
if event.operator == schemas.ClickEventExtraOperator._on_selector:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.selector = %({e_k})s", event.value, value_key=e_k))
|
|
else:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.{events.EventType.CLICK.column} {op} %({e_k})s", event.value,
|
|
value_key=e_k))
|
|
else:
|
|
event_from = event_from % f"{events.EventType.CLICK_IOS.table} AS main "
|
|
if not is_any:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.{events.EventType.CLICK_IOS.column} {op} %({e_k})s", event.value,
|
|
value_key=e_k))
|
|
|
|
elif event_type == events.EventType.INPUT.ui_type:
|
|
if platform == "web":
|
|
event_from = event_from % f"{events.EventType.INPUT.table} AS main "
|
|
if not is_any:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.{events.EventType.INPUT.column} {op} %({e_k})s", event.value,
|
|
value_key=e_k))
|
|
if event.source is not None and len(event.source) > 0:
|
|
event_where.append(sh.multi_conditions(f"main.value ILIKE %(custom{i})s", event.source,
|
|
value_key=f"custom{i}"))
|
|
full_args = {**full_args, **sh.multi_values(event.source, value_key=f"custom{i}")}
|
|
|
|
else:
|
|
event_from = event_from % f"{events.EventType.INPUT_IOS.table} AS main "
|
|
if not is_any:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.{events.EventType.INPUT_IOS.column} {op} %({e_k})s", event.value,
|
|
value_key=e_k))
|
|
|
|
|
|
elif event_type == events.EventType.LOCATION.ui_type:
|
|
if platform == "web":
|
|
event_from = event_from % f"{events.EventType.LOCATION.table} AS main "
|
|
if not is_any:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.{events.EventType.LOCATION.column} {op} %({e_k})s",
|
|
event.value, value_key=e_k))
|
|
else:
|
|
event_from = event_from % f"{events.EventType.VIEW_IOS.table} AS main "
|
|
if not is_any:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.{events.EventType.VIEW_IOS.column} {op} %({e_k})s",
|
|
event.value, value_key=e_k))
|
|
elif event_type == events.EventType.SWIPE_IOS.ui_type and platform == "ios":
|
|
event_from = event_from % f"{events.EventType.SWIPE_IOS.table} AS main "
|
|
if not is_any:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.{events.EventType.SWIPE_IOS.column} {op} %({e_k})s",
|
|
event.value, value_key=e_k))
|
|
elif event_type == events.EventType.CUSTOM.ui_type:
|
|
event_from = event_from % f"{events.EventType.CUSTOM.table} AS main "
|
|
if not is_any:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.{events.EventType.CUSTOM.column} {op} %({e_k})s", event.value,
|
|
value_key=e_k))
|
|
elif event_type == events.EventType.REQUEST.ui_type:
|
|
event_from = event_from % f"{events.EventType.REQUEST.table} AS main "
|
|
if not is_any:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.{events.EventType.REQUEST.column} {op} %({e_k})s", event.value,
|
|
value_key=e_k))
|
|
# elif event_type == events.event_type.GRAPHQL.ui_type:
|
|
# event_from = event_from % f"{events.event_type.GRAPHQL.table} AS main "
|
|
# if not is_any:
|
|
# event_where.append(
|
|
# _multiple_conditions(f"main.{events.event_type.GRAPHQL.column} {op} %({e_k})s", event.value,
|
|
# value_key=e_k))
|
|
elif event_type == events.EventType.STATEACTION.ui_type:
|
|
event_from = event_from % f"{events.EventType.STATEACTION.table} AS main "
|
|
if not is_any:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.{events.EventType.STATEACTION.column} {op} %({e_k})s",
|
|
event.value, value_key=e_k))
|
|
elif event_type == events.EventType.ERROR.ui_type:
|
|
event_from = event_from % f"{events.EventType.ERROR.table} AS main INNER JOIN public.errors AS main1 USING(error_id)"
|
|
event.source = list(set(event.source))
|
|
if not is_any and event.value not in [None, "*", ""]:
|
|
event_where.append(
|
|
sh.multi_conditions(f"(main1.message {op} %({e_k})s OR main1.name {op} %({e_k})s)",
|
|
event.value, value_key=e_k))
|
|
if len(event.source) > 0 and event.source[0] not in [None, "*", ""]:
|
|
event_where.append(sh.multi_conditions(f"main1.source = %({s_k})s", event.source, value_key=s_k))
|
|
|
|
|
|
# ----- IOS
|
|
elif event_type == events.EventType.CLICK_IOS.ui_type:
|
|
event_from = event_from % f"{events.EventType.CLICK_IOS.table} AS main "
|
|
if not is_any:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.{events.EventType.CLICK_IOS.column} {op} %({e_k})s",
|
|
event.value, value_key=e_k))
|
|
|
|
elif event_type == events.EventType.INPUT_IOS.ui_type:
|
|
event_from = event_from % f"{events.EventType.INPUT_IOS.table} AS main "
|
|
if not is_any:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.{events.EventType.INPUT_IOS.column} {op} %({e_k})s",
|
|
event.value, value_key=e_k))
|
|
if event.source is not None and len(event.source) > 0:
|
|
event_where.append(sh.multi_conditions(f"main.value ILIKE %(custom{i})s", event.source,
|
|
value_key="custom{i}"))
|
|
full_args = {**full_args, **sh.multi_values(event.source, f"custom{i}")}
|
|
elif event_type == events.EventType.VIEW_IOS.ui_type:
|
|
event_from = event_from % f"{events.EventType.VIEW_IOS.table} AS main "
|
|
if not is_any:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.{events.EventType.VIEW_IOS.column} {op} %({e_k})s",
|
|
event.value, value_key=e_k))
|
|
elif event_type == events.EventType.CUSTOM_IOS.ui_type:
|
|
event_from = event_from % f"{events.EventType.CUSTOM_IOS.table} AS main "
|
|
if not is_any:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.{events.EventType.CUSTOM_IOS.column} {op} %({e_k})s",
|
|
event.value, value_key=e_k))
|
|
elif event_type == events.EventType.REQUEST_IOS.ui_type:
|
|
event_from = event_from % f"{events.EventType.REQUEST_IOS.table} AS main "
|
|
if not is_any:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.{events.EventType.REQUEST_IOS.column} {op} %({e_k})s",
|
|
event.value, value_key=e_k))
|
|
elif event_type == events.EventType.CRASH_IOS.ui_type:
|
|
event_from = event_from % f"{events.EventType.CRASH_IOS.table} AS main INNER JOIN public.crashes_ios AS main1 USING(crash_id)"
|
|
if not is_any and event.value not in [None, "*", ""]:
|
|
event_where.append(
|
|
sh.multi_conditions(f"(main1.reason {op} %({e_k})s OR main1.name {op} %({e_k})s)",
|
|
event.value, value_key=e_k))
|
|
elif event_type == schemas.PerformanceEventType.fetch_failed:
|
|
event_from = event_from % f"{events.EventType.REQUEST.table} AS main "
|
|
if not is_any:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.{events.EventType.REQUEST.column} {op} %({e_k})s",
|
|
event.value, value_key=e_k))
|
|
col = performance_event.get_col(event_type)
|
|
colname = col["column"]
|
|
event_where.append(f"main.{colname} = FALSE")
|
|
# elif event_type == schemas.PerformanceEventType.fetch_duration:
|
|
# event_from = event_from % f"{events.event_type.REQUEST.table} AS main "
|
|
# if not is_any:
|
|
# event_where.append(
|
|
# _multiple_conditions(f"main.{events.event_type.REQUEST.column} {op} %({e_k})s",
|
|
# event.value, value_key=e_k))
|
|
# col = performance_event.get_col(event_type)
|
|
# colname = col["column"]
|
|
# tname = "main"
|
|
# e_k += "_custom"
|
|
# full_args = {**full_args, **_ sh.multiple_values(event.source, value_key=e_k)}
|
|
# event_where.append(f"{tname}.{colname} IS NOT NULL AND {tname}.{colname}>0 AND " +
|
|
# _multiple_conditions(f"{tname}.{colname} {event.sourceOperator} %({e_k})s",
|
|
# event.source, value_key=e_k))
|
|
elif event_type in [schemas.PerformanceEventType.location_dom_complete,
|
|
schemas.PerformanceEventType.location_largest_contentful_paint_time,
|
|
schemas.PerformanceEventType.location_ttfb,
|
|
schemas.PerformanceEventType.location_avg_cpu_load,
|
|
schemas.PerformanceEventType.location_avg_memory_usage
|
|
]:
|
|
event_from = event_from % f"{events.EventType.LOCATION.table} AS main "
|
|
col = performance_event.get_col(event_type)
|
|
colname = col["column"]
|
|
tname = "main"
|
|
if col.get("extraJoin") is not None:
|
|
tname = "ej"
|
|
event_from += f" INNER JOIN {col['extraJoin']} AS {tname} USING(session_id)"
|
|
event_where += [f"{tname}.timestamp >= main.timestamp", f"{tname}.timestamp >= %(startDate)s",
|
|
f"{tname}.timestamp <= %(endDate)s"]
|
|
if not is_any:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.{events.EventType.LOCATION.column} {op} %({e_k})s",
|
|
event.value, value_key=e_k))
|
|
e_k += "_custom"
|
|
full_args = {**full_args, **sh.multi_values(event.source, value_key=e_k)}
|
|
|
|
event_where.append(f"{tname}.{colname} IS NOT NULL AND {tname}.{colname}>0 AND " +
|
|
sh.multi_conditions(f"{tname}.{colname} {event.sourceOperator.value} %({e_k})s",
|
|
event.source, value_key=e_k))
|
|
|
|
elif event_type == schemas.EventType.request_details:
|
|
event_from = event_from % f"{events.EventType.REQUEST.table} AS main "
|
|
apply = False
|
|
for j, f in enumerate(event.filters):
|
|
is_any = sh.isAny_opreator(f.operator)
|
|
if is_any or len(f.value) == 0:
|
|
continue
|
|
f.value = helper.values_for_operator(value=f.value, op=f.operator)
|
|
op = sh.get_sql_operator(f.operator)
|
|
e_k_f = e_k + f"_fetch{j}"
|
|
full_args = {**full_args, **sh.multi_values(f.value, value_key=e_k_f)}
|
|
if f.type == schemas.FetchFilterType._url:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.{events.EventType.REQUEST.column} {op} %({e_k_f})s::text",
|
|
f.value, value_key=e_k_f))
|
|
apply = True
|
|
elif f.type == schemas.FetchFilterType._status_code:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.status_code {f.operator} %({e_k_f})s::integer", f.value,
|
|
value_key=e_k_f))
|
|
apply = True
|
|
elif f.type == schemas.FetchFilterType._method:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.method {op} %({e_k_f})s", f.value, value_key=e_k_f))
|
|
apply = True
|
|
elif f.type == schemas.FetchFilterType._duration:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.duration {f.operator} %({e_k_f})s::integer", f.value,
|
|
value_key=e_k_f))
|
|
apply = True
|
|
elif f.type == schemas.FetchFilterType._request_body:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.request_body {op} %({e_k_f})s::text", f.value,
|
|
value_key=e_k_f))
|
|
apply = True
|
|
elif f.type == schemas.FetchFilterType._response_body:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.response_body {op} %({e_k_f})s::text", f.value,
|
|
value_key=e_k_f))
|
|
apply = True
|
|
else:
|
|
logging.warning(f"undefined FETCH filter: {f.type}")
|
|
if not apply:
|
|
continue
|
|
elif event_type == schemas.EventType.graphql:
|
|
event_from = event_from % f"{events.EventType.GRAPHQL.table} AS main "
|
|
for j, f in enumerate(event.filters):
|
|
is_any = sh.isAny_opreator(f.operator)
|
|
if is_any or len(f.value) == 0:
|
|
continue
|
|
f.value = helper.values_for_operator(value=f.value, op=f.operator)
|
|
op = sh.get_sql_operator(f.operator)
|
|
e_k_f = e_k + f"_graphql{j}"
|
|
full_args = {**full_args, **sh.multi_values(f.value, value_key=e_k_f)}
|
|
if f.type == schemas.GraphqlFilterType._name:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.{events.EventType.GRAPHQL.column} {op} %({e_k_f})s", f.value,
|
|
value_key=e_k_f))
|
|
elif f.type == schemas.GraphqlFilterType._method:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.method {op} %({e_k_f})s", f.value, value_key=e_k_f))
|
|
elif f.type == schemas.GraphqlFilterType._request_body:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.request_body {op} %({e_k_f})s", f.value, value_key=e_k_f))
|
|
elif f.type == schemas.GraphqlFilterType._response_body:
|
|
event_where.append(
|
|
sh.multi_conditions(f"main.response_body {op} %({e_k_f})s", f.value, value_key=e_k_f))
|
|
else:
|
|
logging.warning(f"undefined GRAPHQL filter: {f.type}")
|
|
else:
|
|
continue
|
|
if event_index == 0 or or_events:
|
|
event_where += ss_constraints
|
|
if is_not:
|
|
if event_index == 0 or or_events:
|
|
events_query_from.append(f"""\
|
|
(SELECT
|
|
session_id,
|
|
0 AS timestamp
|
|
FROM sessions
|
|
WHERE EXISTS(SELECT session_id
|
|
FROM {event_from}
|
|
WHERE {" AND ".join(event_where)}
|
|
AND sessions.session_id=ms.session_id) IS FALSE
|
|
AND project_id = %(projectId)s
|
|
AND start_ts >= %(startDate)s
|
|
AND start_ts <= %(endDate)s
|
|
AND duration IS NOT NULL
|
|
) {"" if or_events else (f"AS event_{event_index}" + ("ON(TRUE)" if event_index > 0 else ""))}\
|
|
""")
|
|
else:
|
|
events_query_from.append(f"""\
|
|
(SELECT
|
|
event_0.session_id,
|
|
event_{event_index - 1}.timestamp AS timestamp
|
|
WHERE EXISTS(SELECT session_id FROM {event_from} WHERE {" AND ".join(event_where)}) IS FALSE
|
|
) AS event_{event_index} {"ON(TRUE)" if event_index > 0 else ""}\
|
|
""")
|
|
else:
|
|
events_query_from.append(f"""\
|
|
(SELECT main.session_id, {"MIN" if event_index < (valid_events_count - 1) else "MAX"}(main.timestamp) AS timestamp
|
|
FROM {event_from}
|
|
WHERE {" AND ".join(event_where)}
|
|
GROUP BY 1
|
|
) {"" if or_events else (f"AS event_{event_index} " + ("ON(TRUE)" if event_index > 0 else ""))}\
|
|
""")
|
|
event_index += 1
|
|
if event_index > 0:
|
|
if or_events:
|
|
events_query_part = f"""SELECT
|
|
session_id,
|
|
MIN(timestamp) AS first_event_ts,
|
|
MAX(timestamp) AS last_event_ts
|
|
FROM ({events_joiner.join(events_query_from)}) AS u
|
|
GROUP BY 1"""
|
|
else:
|
|
events_query_part = f"""SELECT
|
|
event_0.session_id,
|
|
MIN(event_0.timestamp) AS first_event_ts,
|
|
MAX(event_{event_index - 1}.timestamp) AS last_event_ts
|
|
FROM {events_joiner.join(events_query_from)}
|
|
GROUP BY 1"""
|
|
else:
|
|
data.events = []
|
|
# ---------------------------------------------------------------------------
|
|
if data.startTimestamp is not None:
|
|
extra_constraints.append("s.start_ts >= %(startDate)s")
|
|
if data.endTimestamp is not None:
|
|
extra_constraints.append("s.start_ts <= %(endDate)s")
|
|
# if data.platform is not None:
|
|
# if data.platform == schemas.PlatformType.mobile:
|
|
# extra_constraints.append(b"s.user_os in ('Android','BlackBerry OS','iOS','Tizen','Windows Phone')")
|
|
# elif data.platform == schemas.PlatformType.desktop:
|
|
# extra_constraints.append(
|
|
# b"s.user_os in ('Chrome OS','Fedora','Firefox OS','Linux','Mac OS X','Ubuntu','Windows')")
|
|
|
|
if errors_only:
|
|
extra_from += f" INNER JOIN {events.EventType.ERROR.table} AS er USING (session_id) INNER JOIN public.errors AS ser USING (error_id)"
|
|
extra_constraints.append("ser.source = 'js_exception'")
|
|
extra_constraints.append("ser.project_id = %(project_id)s")
|
|
# if error_status != schemas.ErrorStatus.all:
|
|
# extra_constraints.append("ser.status = %(error_status)s")
|
|
# full_args["error_status"] = error_status
|
|
# if favorite_only:
|
|
# extra_from += " INNER JOIN public.user_favorite_errors AS ufe USING (error_id)"
|
|
# extra_constraints.append("ufe.user_id = %(userId)s")
|
|
|
|
if favorite_only and not errors_only and user_id is not None:
|
|
extra_from += """INNER JOIN (SELECT user_id, session_id
|
|
FROM public.user_favorite_sessions
|
|
WHERE user_id = %(userId)s) AS favorite_sessions
|
|
USING (session_id)"""
|
|
elif not favorite_only and not errors_only and user_id is not None:
|
|
extra_from += """LEFT JOIN (SELECT user_id, session_id
|
|
FROM public.user_favorite_sessions
|
|
WHERE user_id = %(userId)s) AS favorite_sessions
|
|
USING (session_id)"""
|
|
extra_join = ""
|
|
if issue is not None:
|
|
extra_join = """
|
|
INNER JOIN LATERAL(SELECT TRUE FROM events_common.issues INNER JOIN public.issues AS p_issues USING (issue_id)
|
|
WHERE issues.session_id=f.session_id
|
|
AND p_issues.type=%(issue_type)s
|
|
AND p_issues.context_string=%(issue_contextString)s
|
|
AND timestamp >= f.first_event_ts
|
|
AND timestamp <= f.last_event_ts) AS issues ON(TRUE)
|
|
"""
|
|
full_args["issue_contextString"] = issue["contextString"]
|
|
full_args["issue_type"] = issue["type"]
|
|
elif len(issues) > 0:
|
|
issues_conditions = []
|
|
for i, f in enumerate(issues):
|
|
f_k_v = f"f_issue_v{i}"
|
|
f_k_s = f_k_v + "_source"
|
|
full_args = {**full_args, **sh.multi_values(f.value, value_key=f_k_v), f_k_s: f.source}
|
|
issues_conditions.append(sh.multi_conditions(f"p_issues.type=%({f_k_v})s", f.value,
|
|
value_key=f_k_v))
|
|
issues_conditions[-1] = f"({issues_conditions[-1]} AND p_issues.context_string=%({f_k_s})s)"
|
|
|
|
if len(events_query_part) > 0:
|
|
extra_join = f"""
|
|
INNER JOIN LATERAL(SELECT TRUE FROM events_common.issues INNER JOIN public.issues AS p_issues USING (issue_id)
|
|
WHERE issues.session_id=f.session_id
|
|
AND timestamp >= f.first_event_ts
|
|
AND timestamp <= f.last_event_ts
|
|
AND {" OR ".join(issues_conditions)}
|
|
) AS issues ON(TRUE)
|
|
"""
|
|
else:
|
|
extra_join = f"""
|
|
INNER JOIN LATERAL(SELECT TRUE FROM events_common.issues INNER JOIN public.issues AS p_issues USING (issue_id)
|
|
WHERE issues.session_id=s.session_id
|
|
AND timestamp >= %(startDate)s
|
|
AND timestamp <= %(endDate)s
|
|
AND {" OR ".join(issues_conditions)}
|
|
) AS issues ON(TRUE)
|
|
"""
|
|
# full_args["issue_contextString"] = issue["contextString"]
|
|
# full_args["issue_type"] = issue["type"]
|
|
if extra_event:
|
|
extra_join += f"""INNER JOIN {extra_event} AS ev USING(session_id)"""
|
|
extra_constraints.append("ev.timestamp>=%(startDate)s")
|
|
extra_constraints.append("ev.timestamp<=%(endDate)s")
|
|
query_part = f"""\
|
|
FROM {f"({events_query_part}) AS f" if len(events_query_part) > 0 else "public.sessions AS s"}
|
|
{extra_join}
|
|
{"INNER JOIN public.sessions AS s USING(session_id)" if len(events_query_part) > 0 else ""}
|
|
{extra_from}
|
|
WHERE
|
|
{" AND ".join(extra_constraints)}"""
|
|
return full_args, query_part
|
|
|
|
|
|
def search_by_metadata(tenant_id, user_id, m_key, m_value, project_id=None):
|
|
if project_id is None:
|
|
all_projects = projects.get_projects(tenant_id=tenant_id)
|
|
else:
|
|
all_projects = [
|
|
projects.get_project(tenant_id=tenant_id, project_id=int(project_id), include_last_session=False,
|
|
include_gdpr=False)]
|
|
|
|
all_projects = {int(p["projectId"]): p["name"] for p in all_projects}
|
|
project_ids = list(all_projects.keys())
|
|
|
|
available_keys = metadata.get_keys_by_projects(project_ids)
|
|
for i in available_keys:
|
|
available_keys[i]["user_id"] = schemas.FilterType.user_id
|
|
available_keys[i]["user_anonymous_id"] = schemas.FilterType.user_anonymous_id
|
|
results = {}
|
|
for i in project_ids:
|
|
if m_key not in available_keys[i].values():
|
|
available_keys.pop(i)
|
|
results[i] = {"total": 0, "sessions": [], "missingMetadata": True}
|
|
project_ids = list(available_keys.keys())
|
|
if len(project_ids) > 0:
|
|
with pg_client.PostgresClient() as cur:
|
|
sub_queries = []
|
|
for i in project_ids:
|
|
col_name = list(available_keys[i].keys())[list(available_keys[i].values()).index(m_key)]
|
|
sub_queries.append(cur.mogrify(
|
|
f"(SELECT COALESCE(COUNT(s.*)) AS count FROM public.sessions AS s WHERE s.project_id = %(id)s AND s.{col_name} = %(value)s) AS \"{i}\"",
|
|
{"id": i, "value": m_value}).decode('UTF-8'))
|
|
query = f"""SELECT {", ".join(sub_queries)};"""
|
|
cur.execute(query=query)
|
|
|
|
rows = cur.fetchone()
|
|
|
|
sub_queries = []
|
|
for i in rows.keys():
|
|
results[i] = {"total": rows[i], "sessions": [], "missingMetadata": False, "name": all_projects[int(i)]}
|
|
if rows[i] > 0:
|
|
col_name = list(available_keys[int(i)].keys())[list(available_keys[int(i)].values()).index(m_key)]
|
|
sub_queries.append(
|
|
cur.mogrify(
|
|
f"""(
|
|
SELECT *
|
|
FROM (
|
|
SELECT DISTINCT ON(favorite_sessions.session_id, s.session_id) {SESSION_PROJECTION_COLS}
|
|
FROM public.sessions AS s LEFT JOIN (SELECT session_id
|
|
FROM public.user_favorite_sessions
|
|
WHERE user_favorite_sessions.user_id = %(userId)s
|
|
) AS favorite_sessions USING (session_id)
|
|
WHERE s.project_id = %(id)s AND s.duration IS NOT NULL AND s.{col_name} = %(value)s
|
|
) AS full_sessions
|
|
ORDER BY favorite DESC, issue_score DESC
|
|
LIMIT 10
|
|
)""",
|
|
{"id": i, "value": m_value, "userId": user_id}).decode('UTF-8'))
|
|
if len(sub_queries) > 0:
|
|
cur.execute("\nUNION\n".join(sub_queries))
|
|
rows = cur.fetchall()
|
|
for i in rows:
|
|
results[str(i["project_id"])]["sessions"].append(helper.dict_to_camel_case(i))
|
|
return results
|
|
|
|
|
|
def get_user_sessions(project_id, user_id, start_date, end_date):
|
|
with pg_client.PostgresClient() as cur:
|
|
constraints = ["s.project_id = %(projectId)s", "s.user_id = %(userId)s"]
|
|
if start_date is not None:
|
|
constraints.append("s.start_ts >= %(startDate)s")
|
|
if end_date is not None:
|
|
constraints.append("s.start_ts <= %(endDate)s")
|
|
|
|
query_part = f"""\
|
|
FROM public.sessions AS s
|
|
WHERE {" AND ".join(constraints)}"""
|
|
|
|
cur.execute(cur.mogrify(f"""\
|
|
SELECT s.project_id,
|
|
s.session_id::text AS session_id,
|
|
s.user_uuid,
|
|
s.user_id,
|
|
s.user_os,
|
|
s.user_browser,
|
|
s.user_device,
|
|
s.user_country,
|
|
s.start_ts,
|
|
s.duration,
|
|
s.events_count,
|
|
s.pages_count,
|
|
s.errors_count
|
|
{query_part}
|
|
ORDER BY s.session_id
|
|
LIMIT 50;""", {
|
|
"projectId": project_id,
|
|
"userId": user_id,
|
|
"startDate": start_date,
|
|
"endDate": end_date
|
|
}))
|
|
|
|
sessions = cur.fetchall()
|
|
return helper.list_to_camel_case(sessions)
|
|
|
|
|
|
def get_session_user(project_id, user_id):
|
|
with pg_client.PostgresClient() as cur:
|
|
query = cur.mogrify(
|
|
"""\
|
|
SELECT
|
|
user_id,
|
|
count(*) as session_count,
|
|
max(start_ts) as last_seen,
|
|
min(start_ts) as first_seen
|
|
FROM
|
|
"public".sessions
|
|
WHERE
|
|
project_id = %(project_id)s
|
|
AND user_id = %(userId)s
|
|
AND duration is not null
|
|
GROUP BY user_id;
|
|
""",
|
|
{"project_id": project_id, "userId": user_id}
|
|
)
|
|
cur.execute(query=query)
|
|
data = cur.fetchone()
|
|
return helper.dict_to_camel_case(data)
|
|
|
|
|
|
def count_all():
|
|
with pg_client.PostgresClient(unlimited_query=True) as cur:
|
|
cur.execute(query="SELECT COUNT(session_id) AS count FROM public.sessions")
|
|
row = cur.fetchone()
|
|
return row.get("count", 0) if row else 0
|
|
|
|
|
|
def session_exists(project_id, session_id):
|
|
with pg_client.PostgresClient() as cur:
|
|
query = cur.mogrify("""SELECT 1
|
|
FROM public.sessions
|
|
WHERE session_id=%(session_id)s
|
|
AND project_id=%(project_id)s
|
|
LIMIT 1;""",
|
|
{"project_id": project_id, "session_id": session_id})
|
|
cur.execute(query)
|
|
row = cur.fetchone()
|
|
return row is not None
|
|
|
|
|
|
def check_recording_status(project_id: int) -> dict:
|
|
query = f"""
|
|
WITH project_sessions AS (SELECT COUNT(1) AS full_count,
|
|
COUNT(1) FILTER ( WHERE duration IS NOT NULL) AS nn_duration_count
|
|
FROM public.sessions
|
|
WHERE project_id = %(project_id)s
|
|
AND start_ts >= (extract(EPOCH FROM now() - INTERVAL '1 day')) * 1000
|
|
AND start_ts <= (extract(EPOCH FROM now() + INTERVAL '1 day')) * 1000)
|
|
SELECT CASE
|
|
WHEN full_count = 0 THEN 0
|
|
WHEN nn_duration_count = 0 THEN 1
|
|
ELSE 2
|
|
END AS recording_status,
|
|
full_count AS sessions_count
|
|
FROM project_sessions;
|
|
"""
|
|
|
|
with pg_client.PostgresClient() as cur:
|
|
query = cur.mogrify(query, {"project_id": project_id})
|
|
cur.execute(query)
|
|
row = cur.fetchone()
|
|
|
|
return {
|
|
"recordingStatus": row["recording_status"],
|
|
"sessionsCount": row["sessions_count"]
|
|
}
|
|
|
|
|
|
def search_sessions_by_ids(project_id: int, session_ids: list, sort_by: str = 'session_id',
|
|
ascending: bool = False) -> dict:
|
|
if session_ids is None or len(session_ids) == 0:
|
|
return {"total": 0, "sessions": []}
|
|
with pg_client.PostgresClient() as cur:
|
|
meta_keys = metadata.get(project_id=project_id)
|
|
params = {"project_id": project_id, "session_ids": tuple(session_ids)}
|
|
order_direction = 'ASC' if ascending else 'DESC'
|
|
main_query = cur.mogrify(f"""SELECT {SESSION_PROJECTION_BASE_COLS}
|
|
{"," if len(meta_keys) > 0 else ""}{",".join([f'metadata_{m["index"]}' for m in meta_keys])}
|
|
FROM public.sessions AS s
|
|
WHERE project_id=%(project_id)s
|
|
AND session_id IN %(session_ids)s
|
|
ORDER BY {sort_by} {order_direction};""", params)
|
|
|
|
cur.execute(main_query)
|
|
rows = cur.fetchall()
|
|
if len(meta_keys) > 0:
|
|
for s in rows:
|
|
s["metadata"] = {}
|
|
for m in meta_keys:
|
|
s["metadata"][m["key"]] = s.pop(f'metadata_{m["index"]}')
|
|
return {"total": len(rows), "sessions": helper.list_to_camel_case(rows)}
|