Compare commits
21 commits
main
...
kafka-topi
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
dda7affd5d | ||
|
|
6e2a772e7f | ||
|
|
dcce3569fb | ||
|
|
867247dbc0 | ||
|
|
754293e29d | ||
|
|
ddd037ce79 | ||
|
|
66f4c5c93b | ||
|
|
66e4d133ad | ||
|
|
f9f8853ab0 | ||
|
|
e0bb6fea9d | ||
|
|
4e7efaecde | ||
|
|
54a9624332 | ||
|
|
1ddffca572 | ||
|
|
c91881413a | ||
|
|
ba2d9eb81c | ||
|
|
c845415e1e | ||
|
|
ee0ede8478 | ||
|
|
72afae226b | ||
|
|
b3f545849a | ||
|
|
cd2966fb9f | ||
|
|
4b91dcded0 |
41 changed files with 457 additions and 245 deletions
10
README_AR.md
10
README_AR.md
|
|
@ -55,17 +55,17 @@ OpenReplay هو مجموعة إعادة تشغيل الجلسة التي يمك
|
||||||
## الميزات
|
## الميزات
|
||||||
|
|
||||||
- **إعادة تشغيل الجلسة:** تتيح لك إعادة تشغيل الجلسة إعادة عيش تجربة مستخدميك، ورؤية أين يواجهون صعوبة وكيف يؤثر ذلك على سلوكهم. يتم تحليل كل إعادة تشغيل للجلسة تلقائيًا بناءً على الأساليب الاستدلالية، لسهولة التقييم.
|
- **إعادة تشغيل الجلسة:** تتيح لك إعادة تشغيل الجلسة إعادة عيش تجربة مستخدميك، ورؤية أين يواجهون صعوبة وكيف يؤثر ذلك على سلوكهم. يتم تحليل كل إعادة تشغيل للجلسة تلقائيًا بناءً على الأساليب الاستدلالية، لسهولة التقييم.
|
||||||
- **أدوات التطوير (DevTools):** إنها مثل التصحيح في متصفحك الخاص. يوفر لك OpenReplay السياق الكامل (نشاط الشبكة، أخطاء JavaScript، إجراءات/حالة التخزين وأكثر من 40 مقياسًا) حتى تتمكن من إعادة إنتاج الأخطاء فورًا وفهم مشكلات الأداء.
|
- **أدوات التطوير (DevTools):** إنها مثل المصحح (debugger) في متصفحك الخاص. يوفر لك OpenReplay السياق الكامل (نشاط الشبكة، أخطاء JavaScript، إجراءات/حالة التخزين وأكثر من 40 مقياسًا) حتى تتمكن من إعادة إنتاج الأخطاء فورًا وفهم مشكلات الأداء.
|
||||||
- **المساعدة (Assist):** تساعدك في دعم مستخدميك من خلال رؤية شاشتهم مباشرة والانضمام فورًا إلى مكالمة (WebRTC) معهم دون الحاجة إلى برامج مشاركة الشاشة من جهات خارجية.
|
- **المساعدة (Assist):** تساعدك في دعم مستخدميك من خلال رؤية شاشتهم مباشرة والانضمام فورًا إلى مكالمة (WebRTC) معهم دون الحاجة إلى برامج مشاركة الشاشة من جهات خارجية.
|
||||||
- **البحث الشامل (Omni-search):** ابحث وفرز حسب أي عملية/معيار للمستخدم تقريبًا، أو سمة الجلسة أو الحدث التقني، حتى تتمكن من الرد على أي سؤال. لا يلزم تجهيز.
|
- **البحث الشامل (Omni-search):** ابحث وافرز حسب أي عملية/معيار للمستخدم تقريبًا، أو سمة الجلسة أو الحدث التقني، حتى تتمكن من الرد على أي سؤال. لا يلزم تجهيز.
|
||||||
- **الأنفاق (Funnels):** للكشف عن المشكلات الأكثر تأثيرًا التي تسبب في فقدان التحويل والإيرادات.
|
- **الأنفاق (Funnels):** للكشف عن المشكلات الأكثر تأثيرًا التي تسبب في فقدان التحويل والإيرادات.
|
||||||
- **ضوابط الخصوصية الدقيقة:** اختر ماذا تريد التقاطه، ماذا تريد أن تخفي أو تجاهل حتى لا تصل بيانات المستخدم حتى إلى خوادمك.
|
- **ضوابط الخصوصية الدقيقة:** اختر ماذا تريد التقاطه، ماذا تريد أن تخفي أو تتجاهل حتى لا تصل بيانات المستخدم حتى إلى خوادمك.
|
||||||
- **موجهة للمكونات الإضافية (Plugins oriented):** تصل إلى السبب الجذري بشكل أسرع عن طريق تتبع حالة التطبيق (Redux، VueX، MobX، NgRx، Pinia، وZustand) وتسجيل استعلامات GraphQL (Apollo، Relay) وطلبات Fetch/Axios.
|
- **موجهة للمكونات الإضافية (Plugins oriented):** يمكنك الوصول إلى السبب الجذري بشكل أسرع عن طريق تتبع حالة التطبيق (Redux، VueX، MobX، NgRx، Pinia، وZustand) وتسجيل استعلامات GraphQL (Apollo، Relay) وطلبات Fetch/Axios.
|
||||||
- **التكاملات (Integrations):** مزامنة سجلات الخادم الخلفي مع إعادات التشغيل للجلسات ورؤية ما حدث من الأمام إلى الخلف. يدعم OpenReplay Sentry وDatadog وCloudWatch وStackdriver وElastic والمزيد.
|
- **التكاملات (Integrations):** مزامنة سجلات الخادم الخلفي مع إعادات التشغيل للجلسات ورؤية ما حدث من الأمام إلى الخلف. يدعم OpenReplay Sentry وDatadog وCloudWatch وStackdriver وElastic والمزيد.
|
||||||
|
|
||||||
## خيارات النشر
|
## خيارات النشر
|
||||||
|
|
||||||
يمكن نشر OpenReplay في أي مكان. اتبع دليلنا الخطوة بالخطوة لنشره على خدمات السحابة العامة الرئيسية:
|
يمكن نشر OpenReplay في أي مكان. اتبع دليلنا خطوة بخطوة لنشره على خدمات السحابة العامة الرئيسة:
|
||||||
|
|
||||||
- [AWS](https://docs.openreplay.com/deployment/deploy-aws)
|
- [AWS](https://docs.openreplay.com/deployment/deploy-aws)
|
||||||
- [Google Cloud](https://docs.openreplay.com/deployment/deploy-gcp)
|
- [Google Cloud](https://docs.openreplay.com/deployment/deploy-gcp)
|
||||||
|
|
|
||||||
|
|
@ -319,13 +319,14 @@ def create_card(project_id, user_id, data: schemas.CardSchema, dashboard=False):
|
||||||
session_data = None
|
session_data = None
|
||||||
if data.metric_type == schemas.MetricType.heat_map:
|
if data.metric_type == schemas.MetricType.heat_map:
|
||||||
if data.session_id is not None:
|
if data.session_id is not None:
|
||||||
session_data = json.dumps({"sessionId": data.session_id})
|
session_data = {"sessionId": data.session_id}
|
||||||
else:
|
else:
|
||||||
session_data = __get_heat_map_chart(project_id=project_id, user_id=user_id,
|
session_data = __get_heat_map_chart(project_id=project_id, user_id=user_id,
|
||||||
data=data, include_mobs=False)
|
data=data, include_mobs=False)
|
||||||
if session_data is not None:
|
if session_data is not None:
|
||||||
session_data = json.dumps({"sessionId": session_data["sessionId"]})
|
session_data = {"sessionId": session_data["sessionId"]}
|
||||||
_data = {"session_data": session_data}
|
|
||||||
|
_data = {"session_data": json.dumps(session_data) if session_data is not None else None}
|
||||||
for i, s in enumerate(data.series):
|
for i, s in enumerate(data.series):
|
||||||
for k in s.model_dump().keys():
|
for k in s.model_dump().keys():
|
||||||
_data[f"{k}_{i}"] = s.__getattribute__(k)
|
_data[f"{k}_{i}"] = s.__getattribute__(k)
|
||||||
|
|
|
||||||
|
|
@ -359,12 +359,12 @@ def search2_table(data: schemas.SessionsSearchPayloadSchema, project_id: int, de
|
||||||
distinct_on += ",path"
|
distinct_on += ",path"
|
||||||
if metric_format == schemas.MetricExtendedFormatType.session_count:
|
if metric_format == schemas.MetricExtendedFormatType.session_count:
|
||||||
main_query = f"""SELECT COUNT(*) AS count,
|
main_query = f"""SELECT COUNT(*) AS count,
|
||||||
COALESCE(SUM(users_sessions.session_count),0) AS total_sessions,
|
COALESCE(SUM(users_sessions.session_count),0) AS count,
|
||||||
COALESCE(JSONB_AGG(users_sessions)
|
COALESCE(JSONB_AGG(users_sessions)
|
||||||
FILTER ( WHERE rn > %(limit_s)s
|
FILTER ( WHERE rn > %(limit_s)s
|
||||||
AND rn <= %(limit_e)s ), '[]'::JSONB) AS values
|
AND rn <= %(limit_e)s ), '[]'::JSONB) AS values
|
||||||
FROM (SELECT {main_col} AS name,
|
FROM (SELECT {main_col} AS name,
|
||||||
count(DISTINCT session_id) AS session_count,
|
count(DISTINCT session_id) AS total,
|
||||||
ROW_NUMBER() OVER (ORDER BY count(full_sessions) DESC) AS rn
|
ROW_NUMBER() OVER (ORDER BY count(full_sessions) DESC) AS rn
|
||||||
FROM (SELECT *
|
FROM (SELECT *
|
||||||
FROM (SELECT DISTINCT ON({distinct_on}) s.session_id, s.user_uuid,
|
FROM (SELECT DISTINCT ON({distinct_on}) s.session_id, s.user_uuid,
|
||||||
|
|
@ -379,7 +379,7 @@ def search2_table(data: schemas.SessionsSearchPayloadSchema, project_id: int, de
|
||||||
ORDER BY session_count DESC) AS users_sessions;"""
|
ORDER BY session_count DESC) AS users_sessions;"""
|
||||||
else:
|
else:
|
||||||
main_query = f"""SELECT COUNT(*) AS count,
|
main_query = f"""SELECT COUNT(*) AS count,
|
||||||
COALESCE(SUM(users_sessions.user_count),0) AS total_users,
|
COALESCE(SUM(users_sessions.user_count),0) AS count,
|
||||||
COALESCE(JSONB_AGG(users_sessions) FILTER ( WHERE rn <= 200 ), '[]'::JSONB) AS values
|
COALESCE(JSONB_AGG(users_sessions) FILTER ( WHERE rn <= 200 ), '[]'::JSONB) AS values
|
||||||
FROM (SELECT {main_col} AS name,
|
FROM (SELECT {main_col} AS name,
|
||||||
count(DISTINCT user_id) AS user_count,
|
count(DISTINCT user_id) AS user_count,
|
||||||
|
|
@ -420,12 +420,12 @@ def search_table_of_individual_issues(data: schemas.SessionsSearchPayloadSchema,
|
||||||
full_args["issues_limit_s"] = (data.page - 1) * data.limit
|
full_args["issues_limit_s"] = (data.page - 1) * data.limit
|
||||||
full_args["issues_limit_e"] = data.page * data.limit
|
full_args["issues_limit_e"] = data.page * data.limit
|
||||||
main_query = cur.mogrify(f"""SELECT COUNT(1) AS count,
|
main_query = cur.mogrify(f"""SELECT COUNT(1) AS count,
|
||||||
COALESCE(SUM(session_count), 0) AS total_sessions,
|
COALESCE(SUM(session_count), 0) AS count,
|
||||||
COALESCE(JSONB_AGG(ranked_issues)
|
COALESCE(JSONB_AGG(ranked_issues)
|
||||||
FILTER ( WHERE rn > %(issues_limit_s)s
|
FILTER ( WHERE rn > %(issues_limit_s)s
|
||||||
AND rn <= %(issues_limit_e)s ), '[]'::JSONB) AS values
|
AND rn <= %(issues_limit_e)s ), '[]'::JSONB) AS values
|
||||||
FROM (SELECT *, ROW_NUMBER() OVER (ORDER BY session_count DESC) AS rn
|
FROM (SELECT *, ROW_NUMBER() OVER (ORDER BY session_count DESC) AS rn
|
||||||
FROM (SELECT type AS name, context_string AS value, COUNT(DISTINCT session_id) AS session_count
|
FROM (SELECT type AS name, context_string AS value, COUNT(DISTINCT session_id) AS total
|
||||||
FROM (SELECT session_id
|
FROM (SELECT session_id
|
||||||
{query_part}) AS filtered_sessions
|
{query_part}) AS filtered_sessions
|
||||||
INNER JOIN events_common.issues USING (session_id)
|
INNER JOIN events_common.issues USING (session_id)
|
||||||
|
|
@ -814,12 +814,6 @@ def search_query_parts(data: schemas.SessionsSearchPayloadSchema, error_status,
|
||||||
event_where.append(
|
event_where.append(
|
||||||
sh.multi_conditions(f"main.{events.EventType.VIEW_MOBILE.column} {op} %({e_k})s",
|
sh.multi_conditions(f"main.{events.EventType.VIEW_MOBILE.column} {op} %({e_k})s",
|
||||||
event.value, value_key=e_k))
|
event.value, value_key=e_k))
|
||||||
elif event_type == events.EventType.SWIPE_MOBILE.ui_type and platform == "ios":
|
|
||||||
event_from = event_from % f"{events.EventType.SWIPE_MOBILE.table} AS main "
|
|
||||||
if not is_any:
|
|
||||||
event_where.append(
|
|
||||||
sh.multi_conditions(f"main.{events.EventType.SWIPE_MOBILE.column} {op} %({e_k})s",
|
|
||||||
event.value, value_key=e_k))
|
|
||||||
elif event_type == events.EventType.CUSTOM.ui_type:
|
elif event_type == events.EventType.CUSTOM.ui_type:
|
||||||
event_from = event_from % f"{events.EventType.CUSTOM.table} AS main "
|
event_from = event_from % f"{events.EventType.CUSTOM.table} AS main "
|
||||||
if not is_any:
|
if not is_any:
|
||||||
|
|
@ -855,7 +849,7 @@ def search_query_parts(data: schemas.SessionsSearchPayloadSchema, error_status,
|
||||||
event_where.append(sh.multi_conditions(f"main1.source = %({s_k})s", event.source, value_key=s_k))
|
event_where.append(sh.multi_conditions(f"main1.source = %({s_k})s", event.source, value_key=s_k))
|
||||||
|
|
||||||
|
|
||||||
# ----- IOS
|
# ----- Mobile
|
||||||
elif event_type == events.EventType.CLICK_MOBILE.ui_type:
|
elif event_type == events.EventType.CLICK_MOBILE.ui_type:
|
||||||
event_from = event_from % f"{events.EventType.CLICK_MOBILE.table} AS main "
|
event_from = event_from % f"{events.EventType.CLICK_MOBILE.table} AS main "
|
||||||
if not is_any:
|
if not is_any:
|
||||||
|
|
@ -897,6 +891,13 @@ def search_query_parts(data: schemas.SessionsSearchPayloadSchema, error_status,
|
||||||
event_where.append(
|
event_where.append(
|
||||||
sh.multi_conditions(f"(main1.reason {op} %({e_k})s OR main1.name {op} %({e_k})s)",
|
sh.multi_conditions(f"(main1.reason {op} %({e_k})s OR main1.name {op} %({e_k})s)",
|
||||||
event.value, value_key=e_k))
|
event.value, value_key=e_k))
|
||||||
|
elif event_type == events.EventType.SWIPE_MOBILE.ui_type and platform != "web":
|
||||||
|
event_from = event_from % f"{events.EventType.SWIPE_MOBILE.table} AS main "
|
||||||
|
if not is_any:
|
||||||
|
event_where.append(
|
||||||
|
sh.multi_conditions(f"main.{events.EventType.SWIPE_MOBILE.column} {op} %({e_k})s",
|
||||||
|
event.value, value_key=e_k))
|
||||||
|
|
||||||
elif event_type == schemas.PerformanceEventType.fetch_failed:
|
elif event_type == schemas.PerformanceEventType.fetch_failed:
|
||||||
event_from = event_from % f"{events.EventType.REQUEST.table} AS main "
|
event_from = event_from % f"{events.EventType.REQUEST.table} AS main "
|
||||||
if not is_any:
|
if not is_any:
|
||||||
|
|
|
||||||
|
|
@ -19,6 +19,7 @@ from routers.base import get_routers
|
||||||
public_app, app, app_apikey = get_routers()
|
public_app, app, app_apikey = get_routers()
|
||||||
|
|
||||||
|
|
||||||
|
@app.get('/{projectId}/autocomplete', tags=["events"])
|
||||||
@app.get('/{projectId}/events/search', tags=["events"])
|
@app.get('/{projectId}/events/search', tags=["events"])
|
||||||
def events_search(projectId: int, q: str,
|
def events_search(projectId: int, q: str,
|
||||||
type: Union[schemas.FilterType, schemas.EventType,
|
type: Union[schemas.FilterType, schemas.EventType,
|
||||||
|
|
|
||||||
|
|
@ -471,13 +471,13 @@ class EventType(str, Enum):
|
||||||
state_action = "stateAction"
|
state_action = "stateAction"
|
||||||
error = "error"
|
error = "error"
|
||||||
tag = "tag"
|
tag = "tag"
|
||||||
click_mobile = "tapIos"
|
click_mobile = "click_mobile"
|
||||||
input_mobile = "inputIos"
|
input_mobile = "input_mobile"
|
||||||
view_mobile = "viewIos"
|
view_mobile = "view_mobile"
|
||||||
custom_mobile = "customIos"
|
custom_mobile = "custom_mobile"
|
||||||
request_mobile = "requestIos"
|
request_mobile = "request_mobile"
|
||||||
error_mobile = "errorIos"
|
error_mobile = "error_mobile"
|
||||||
swipe_mobile = "swipeIos"
|
swipe_mobile = "swipe_mobile"
|
||||||
|
|
||||||
|
|
||||||
class PerformanceEventType(str, Enum):
|
class PerformanceEventType(str, Enum):
|
||||||
|
|
|
||||||
|
|
@ -81,13 +81,13 @@ func (s *saverImpl) handleMobileMessage(msg Message) error {
|
||||||
if err = s.sessions.UpdateUserID(session.SessionID, m.ID); err != nil {
|
if err = s.sessions.UpdateUserID(session.SessionID, m.ID); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
s.pg.InsertAutocompleteValue(session.SessionID, session.ProjectID, "USERID_Mobile", m.ID)
|
s.pg.InsertAutocompleteValue(session.SessionID, session.ProjectID, "USERID_MOBILE", m.ID)
|
||||||
return nil
|
return nil
|
||||||
case *MobileUserAnonymousID:
|
case *MobileUserAnonymousID:
|
||||||
if err = s.sessions.UpdateAnonymousID(session.SessionID, m.ID); err != nil {
|
if err = s.sessions.UpdateAnonymousID(session.SessionID, m.ID); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
s.pg.InsertAutocompleteValue(session.SessionID, session.ProjectID, "USERANONYMOUSID_Mobile", m.ID)
|
s.pg.InsertAutocompleteValue(session.SessionID, session.ProjectID, "USERANONYMOUSID_MOBILE", m.ID)
|
||||||
return nil
|
return nil
|
||||||
case *MobileMetadata:
|
case *MobileMetadata:
|
||||||
return s.sessions.UpdateMetadata(m.SessionID(), m.Key, m.Value)
|
return s.sessions.UpdateMetadata(m.SessionID(), m.Key, m.Value)
|
||||||
|
|
|
||||||
|
|
@ -132,8 +132,15 @@ func (conn *Conn) InsertWebClickEvent(sess *sessions.Session, e *messages.MouseC
|
||||||
}
|
}
|
||||||
var host, path string
|
var host, path string
|
||||||
host, path, _, _ = url.GetURLParts(e.Url)
|
host, path, _, _ = url.GetURLParts(e.Url)
|
||||||
if e.NormalizedX <= 100 && e.NormalizedY <= 100 {
|
if e.NormalizedX != 101 && e.NormalizedY != 101 {
|
||||||
if err := conn.bulks.Get("webClickXYEvents").Append(sess.SessionID, truncSqIdx(e.MsgID()), e.Timestamp, e.Label, e.Selector, host+path, path, e.HesitationTime, e.NormalizedX, e.NormalizedY); err != nil {
|
// To support previous versions of tracker
|
||||||
|
if e.NormalizedX <= 100 && e.NormalizedY <= 100 {
|
||||||
|
e.NormalizedX *= 100
|
||||||
|
e.NormalizedY *= 100
|
||||||
|
}
|
||||||
|
normalizedX := float32(e.NormalizedX) / 100.0
|
||||||
|
normalizedY := float32(e.NormalizedY) / 100.0
|
||||||
|
if err := conn.bulks.Get("webClickXYEvents").Append(sess.SessionID, truncSqIdx(e.MsgID()), e.Timestamp, e.Label, e.Selector, host+path, path, e.HesitationTime, normalizedX, normalizedY); err != nil {
|
||||||
sessCtx := context.WithValue(context.Background(), "sessionID", sess.SessionID)
|
sessCtx := context.WithValue(context.Background(), "sessionID", sess.SessionID)
|
||||||
conn.log.Error(sessCtx, "insert web click event in bulk err: %s", err)
|
conn.log.Error(sessCtx, "insert web click event in bulk err: %s", err)
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -13,14 +13,14 @@ func (conn *Conn) InsertMobileEvent(session *sessions.Session, e *messages.Mobil
|
||||||
if err := conn.InsertCustomEvent(session.SessionID, e.Timestamp, truncSqIdx(e.Index), e.Name, e.Payload); err != nil {
|
if err := conn.InsertCustomEvent(session.SessionID, e.Timestamp, truncSqIdx(e.Index), e.Name, e.Payload); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
conn.InsertAutocompleteValue(session.SessionID, session.ProjectID, "CUSTOM_Mobile", e.Name)
|
conn.InsertAutocompleteValue(session.SessionID, session.ProjectID, "CUSTOM_MOBILE", e.Name)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (conn *Conn) InsertMobileNetworkCall(sess *sessions.Session, e *messages.MobileNetworkCall) error {
|
func (conn *Conn) InsertMobileNetworkCall(sess *sessions.Session, e *messages.MobileNetworkCall) error {
|
||||||
err := conn.InsertRequest(sess.SessionID, e.Timestamp, truncSqIdx(e.Index), e.URL, e.Duration, e.Status < 400)
|
err := conn.InsertRequest(sess.SessionID, e.Timestamp, truncSqIdx(e.Index), e.URL, e.Duration, e.Status < 400)
|
||||||
if err == nil {
|
if err == nil {
|
||||||
conn.InsertAutocompleteValue(sess.SessionID, sess.ProjectID, "REQUEST_Mobile", url.DiscardURLQuery(e.URL))
|
conn.InsertAutocompleteValue(sess.SessionID, sess.ProjectID, "REQUEST_MOBILE", url.DiscardURLQuery(e.URL))
|
||||||
}
|
}
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
@ -36,7 +36,7 @@ func (conn *Conn) InsertMobileClickEvent(sess *sessions.Session, clickEvent *mes
|
||||||
); err != nil {
|
); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
conn.InsertAutocompleteValue(sess.SessionID, sess.ProjectID, "CLICK_Mobile", clickEvent.Label)
|
conn.InsertAutocompleteValue(sess.SessionID, sess.ProjectID, "CLICK_MOBILE", clickEvent.Label)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -51,7 +51,7 @@ func (conn *Conn) InsertMobileSwipeEvent(sess *sessions.Session, swipeEvent *mes
|
||||||
); err != nil {
|
); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
conn.InsertAutocompleteValue(sess.SessionID, sess.ProjectID, "SWIPE_Mobile", swipeEvent.Label)
|
conn.InsertAutocompleteValue(sess.SessionID, sess.ProjectID, "SWIPE_MOBILE", swipeEvent.Label)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -66,7 +66,7 @@ func (conn *Conn) InsertMobileInputEvent(sess *sessions.Session, inputEvent *mes
|
||||||
); err != nil {
|
); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
conn.InsertAutocompleteValue(sess.SessionID, sess.ProjectID, "INPUT_Mobile", inputEvent.Label)
|
conn.InsertAutocompleteValue(sess.SessionID, sess.ProjectID, "INPUT_MOBILE", inputEvent.Label)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -29,6 +29,11 @@ def _get_current_auth_context(request: Request, jwt_payload: dict) -> schemas.Cu
|
||||||
return request.state.currentContext
|
return request.state.currentContext
|
||||||
|
|
||||||
|
|
||||||
|
def _allow_access_to_endpoint(request: Request, current_context: schemas.CurrentContext) -> bool:
|
||||||
|
return not current_context.service_account \
|
||||||
|
or request.url.path not in ["/logout", "/api/logout", "/refresh", "/api/refresh"]
|
||||||
|
|
||||||
|
|
||||||
class JWTAuth(HTTPBearer):
|
class JWTAuth(HTTPBearer):
|
||||||
def __init__(self, auto_error: bool = True):
|
def __init__(self, auto_error: bool = True):
|
||||||
super(JWTAuth, self).__init__(auto_error=auto_error)
|
super(JWTAuth, self).__init__(auto_error=auto_error)
|
||||||
|
|
@ -68,7 +73,10 @@ class JWTAuth(HTTPBearer):
|
||||||
or old_jwt_payload.get("userId") != jwt_payload.get("userId"):
|
or old_jwt_payload.get("userId") != jwt_payload.get("userId"):
|
||||||
raise HTTPException(status_code=status.HTTP_403_FORBIDDEN, detail="Invalid token or expired token.")
|
raise HTTPException(status_code=status.HTTP_403_FORBIDDEN, detail="Invalid token or expired token.")
|
||||||
|
|
||||||
return _get_current_auth_context(request=request, jwt_payload=jwt_payload)
|
ctx = _get_current_auth_context(request=request, jwt_payload=jwt_payload)
|
||||||
|
if not _allow_access_to_endpoint(request=request, current_context=ctx):
|
||||||
|
raise HTTPException(status_code=status.HTTP_403_FORBIDDEN, detail="Unauthorized endpoint.")
|
||||||
|
return ctx
|
||||||
|
|
||||||
else:
|
else:
|
||||||
credentials: HTTPAuthorizationCredentials = await super(JWTAuth, self).__call__(request)
|
credentials: HTTPAuthorizationCredentials = await super(JWTAuth, self).__call__(request)
|
||||||
|
|
@ -95,7 +103,10 @@ class JWTAuth(HTTPBearer):
|
||||||
|
|
||||||
raise HTTPException(status_code=status.HTTP_403_FORBIDDEN, detail="Invalid token or expired token.")
|
raise HTTPException(status_code=status.HTTP_403_FORBIDDEN, detail="Invalid token or expired token.")
|
||||||
|
|
||||||
return _get_current_auth_context(request=request, jwt_payload=jwt_payload)
|
ctx = _get_current_auth_context(request=request, jwt_payload=jwt_payload)
|
||||||
|
if not _allow_access_to_endpoint(request=request, current_context=ctx):
|
||||||
|
raise HTTPException(status_code=status.HTTP_403_FORBIDDEN, detail="Unauthorized endpoint.")
|
||||||
|
return ctx
|
||||||
|
|
||||||
logger.warning("Invalid authorization code.")
|
logger.warning("Invalid authorization code.")
|
||||||
raise HTTPException(status_code=status.HTTP_400_BAD_REQUEST, detail="Invalid authorization code.")
|
raise HTTPException(status_code=status.HTTP_400_BAD_REQUEST, detail="Invalid authorization code.")
|
||||||
|
|
|
||||||
|
|
@ -339,10 +339,13 @@ def create_card(project_id, user_id, data: schemas.CardSchema, dashboard=False):
|
||||||
session_data = None
|
session_data = None
|
||||||
if data.metric_type == schemas.MetricType.heat_map:
|
if data.metric_type == schemas.MetricType.heat_map:
|
||||||
if data.session_id is not None:
|
if data.session_id is not None:
|
||||||
session_data = json.dumps({"sessionId": data.session_id})
|
session_data = {"sessionId": data.session_id}
|
||||||
else:
|
else:
|
||||||
session_data = __get_heat_map_chart(project_id=project_id, user_id=user_id,
|
session_data = __get_heat_map_chart(project_id=project_id, user_id=user_id,
|
||||||
data=data, include_mobs=False)
|
data=data, include_mobs=False)
|
||||||
|
if session_data is not None:
|
||||||
|
session_data = {"sessionId": session_data["sessionId"]}
|
||||||
|
|
||||||
if session_data is not None:
|
if session_data is not None:
|
||||||
# for EE only
|
# for EE only
|
||||||
keys = sessions_mobs. \
|
keys = sessions_mobs. \
|
||||||
|
|
@ -356,8 +359,8 @@ def create_card(project_id, user_id, data: schemas.CardSchema, dashboard=False):
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.warning(f"!!!Error while tagging: {k} to {tag} for heatMap")
|
logger.warning(f"!!!Error while tagging: {k} to {tag} for heatMap")
|
||||||
logger.error(str(e))
|
logger.error(str(e))
|
||||||
session_data = json.dumps(session_data)
|
|
||||||
_data = {"session_data": session_data}
|
_data = {"session_data": json.dumps(session_data) if session_data is not None else None}
|
||||||
for i, s in enumerate(data.series):
|
for i, s in enumerate(data.series):
|
||||||
for k in s.model_dump().keys():
|
for k in s.model_dump().keys():
|
||||||
_data[f"{k}_{i}"] = s.__getattribute__(k)
|
_data[f"{k}_{i}"] = s.__getattribute__(k)
|
||||||
|
|
|
||||||
|
|
@ -57,16 +57,16 @@ def get_by_url(project_id, data: schemas.GetHeatMapPayloadSchema):
|
||||||
# f.value, value_key=f_k))
|
# f.value, value_key=f_k))
|
||||||
|
|
||||||
if data.click_rage and not has_click_rage_filter:
|
if data.click_rage and not has_click_rage_filter:
|
||||||
constraints.append("""(issues.session_id IS NULL
|
constraints.append("""(issues_t.session_id IS NULL
|
||||||
OR (issues.datetime >= toDateTime(%(startDate)s/1000)
|
OR (issues_t.datetime >= toDateTime(%(startDate)s/1000)
|
||||||
AND issues.datetime <= toDateTime(%(endDate)s/1000)
|
AND issues_t.datetime <= toDateTime(%(endDate)s/1000)
|
||||||
AND issues.project_id = toUInt16(%(project_id)s)
|
AND issues_t.project_id = toUInt16(%(project_id)s)
|
||||||
AND issues.event_type = 'ISSUE'
|
AND issues_t.event_type = 'ISSUE'
|
||||||
AND issues.project_id = toUInt16(%(project_id)s
|
AND issues_t.project_id = toUInt16(%(project_id)s)
|
||||||
AND mis.project_id = toUInt16(%(project_id)s
|
AND mis.project_id = toUInt16(%(project_id)s)
|
||||||
AND mis.type='click_rage'))))""")
|
AND mis.type='click_rage'))""")
|
||||||
query_from += """ LEFT JOIN experimental.events AS issues ON (main_events.session_id=issues.session_id)
|
query_from += """ LEFT JOIN experimental.events AS issues_t ON (main_events.session_id=issues_t.session_id)
|
||||||
LEFT JOIN experimental.issues AS mis ON (issues.issue_id=mis.issue_id)"""
|
LEFT JOIN experimental.issues AS mis ON (issues_t.issue_id=mis.issue_id)"""
|
||||||
with ch_client.ClickHouseClient() as cur:
|
with ch_client.ClickHouseClient() as cur:
|
||||||
query = cur.format(f"""SELECT main_events.normalized_x AS normalized_x,
|
query = cur.format(f"""SELECT main_events.normalized_x AS normalized_x,
|
||||||
main_events.normalized_y AS normalized_y
|
main_events.normalized_y AS normalized_y
|
||||||
|
|
|
||||||
|
|
@ -442,7 +442,8 @@ def search2_table(data: schemas.SessionsSearchPayloadSchema, project_id: int, de
|
||||||
if metric_format == schemas.MetricExtendedFormatType.session_count:
|
if metric_format == schemas.MetricExtendedFormatType.session_count:
|
||||||
main_query = f"""SELECT COUNT(DISTINCT {main_col}) OVER () AS main_count,
|
main_query = f"""SELECT COUNT(DISTINCT {main_col}) OVER () AS main_count,
|
||||||
{main_col} AS name,
|
{main_col} AS name,
|
||||||
count(DISTINCT session_id) AS session_count
|
count(DISTINCT session_id) AS session_count,
|
||||||
|
COALESCE(SUM(count(DISTINCT session_id)) OVER (), 0) AS total_sessions
|
||||||
FROM (SELECT s.session_id AS session_id,
|
FROM (SELECT s.session_id AS session_id,
|
||||||
{extra_col}
|
{extra_col}
|
||||||
{query_part}) AS filtred_sessions
|
{query_part}) AS filtred_sessions
|
||||||
|
|
@ -470,11 +471,14 @@ def search2_table(data: schemas.SessionsSearchPayloadSchema, project_id: int, de
|
||||||
logging.debug("--------------------")
|
logging.debug("--------------------")
|
||||||
sessions = cur.execute(main_query)
|
sessions = cur.execute(main_query)
|
||||||
count = 0
|
count = 0
|
||||||
|
total_sessions = 0
|
||||||
if len(sessions) > 0:
|
if len(sessions) > 0:
|
||||||
count = sessions[0]["main_count"]
|
count = sessions[0]["main_count"]
|
||||||
|
total_sessions = sessions[0]["total_sessions"]
|
||||||
for s in sessions:
|
for s in sessions:
|
||||||
s.pop("main_count")
|
s.pop("main_count")
|
||||||
sessions = {"count": count, "values": helper.list_to_camel_case(sessions)}
|
s.pop("total_sessions")
|
||||||
|
sessions = {"total": count, "count": total_sessions, "values": helper.list_to_camel_case(sessions)}
|
||||||
|
|
||||||
return sessions
|
return sessions
|
||||||
|
|
||||||
|
|
@ -520,7 +524,7 @@ def search_table_of_individual_issues(data: schemas.SessionsSearchPayloadSchema,
|
||||||
total_sessions = 0
|
total_sessions = 0
|
||||||
issues_count = 0
|
issues_count = 0
|
||||||
|
|
||||||
return {"count": issues_count, "totalSessions": total_sessions, "values": issues}
|
return {"total": issues_count, "count": total_sessions, "values": issues}
|
||||||
|
|
||||||
|
|
||||||
def __is_valid_event(is_any: bool, event: schemas.SessionSearchEventSchema2):
|
def __is_valid_event(is_any: bool, event: schemas.SessionSearchEventSchema2):
|
||||||
|
|
@ -563,7 +567,7 @@ def __get_event_type(event_type: Union[schemas.EventType, schemas.PerformanceEve
|
||||||
schemas.PerformanceEventType.fetch_failed: "REQUEST",
|
schemas.PerformanceEventType.fetch_failed: "REQUEST",
|
||||||
schemas.EventType.error: "CRASH",
|
schemas.EventType.error: "CRASH",
|
||||||
}
|
}
|
||||||
if platform == "ios" and event_type in defs_mobile:
|
if platform != "web" and event_type in defs_mobile:
|
||||||
return defs_mobile.get(event_type)
|
return defs_mobile.get(event_type)
|
||||||
if event_type not in defs:
|
if event_type not in defs:
|
||||||
raise Exception(f"unsupported EventType:{event_type}")
|
raise Exception(f"unsupported EventType:{event_type}")
|
||||||
|
|
@ -964,7 +968,7 @@ def search_query_parts_ch(data: schemas.SessionsSearchPayloadSchema, error_statu
|
||||||
value_key=f"custom{i}"))
|
value_key=f"custom{i}"))
|
||||||
full_args = {**full_args, **_multiple_values(event.source, value_key=f"custom{i}")}
|
full_args = {**full_args, **_multiple_values(event.source, value_key=f"custom{i}")}
|
||||||
else:
|
else:
|
||||||
_column = events.EventType.INPUT_IOS.column
|
_column = events.EventType.INPUT_MOBILE.column
|
||||||
event_where.append(f"main.event_type='{__get_event_type(event_type, platform=platform)}'")
|
event_where.append(f"main.event_type='{__get_event_type(event_type, platform=platform)}'")
|
||||||
events_conditions.append({"type": event_where[-1]})
|
events_conditions.append({"type": event_where[-1]})
|
||||||
if not is_any:
|
if not is_any:
|
||||||
|
|
@ -997,7 +1001,7 @@ def search_query_parts_ch(data: schemas.SessionsSearchPayloadSchema, error_statu
|
||||||
event.value, value_key=e_k))
|
event.value, value_key=e_k))
|
||||||
events_conditions[-1]["condition"] = event_where[-1]
|
events_conditions[-1]["condition"] = event_where[-1]
|
||||||
else:
|
else:
|
||||||
_column = events.EventType.VIEW_IOS.column
|
_column = events.EventType.VIEW_MOBILE.column
|
||||||
event_where.append(f"main.event_type='{__get_event_type(event_type, platform=platform)}'")
|
event_where.append(f"main.event_type='{__get_event_type(event_type, platform=platform)}'")
|
||||||
events_conditions.append({"type": event_where[-1]})
|
events_conditions.append({"type": event_where[-1]})
|
||||||
if not is_any:
|
if not is_any:
|
||||||
|
|
@ -1089,6 +1093,114 @@ def search_query_parts_ch(data: schemas.SessionsSearchPayloadSchema, error_statu
|
||||||
|
|
||||||
events_conditions[-1]["condition"] = " AND ".join(events_conditions[-1]["condition"])
|
events_conditions[-1]["condition"] = " AND ".join(events_conditions[-1]["condition"])
|
||||||
|
|
||||||
|
# ----- Mobile
|
||||||
|
elif event_type == events.EventType.CLICK_MOBILE.ui_type:
|
||||||
|
_column = events.EventType.CLICK_MOBILE.column
|
||||||
|
event_where.append(f"main.event_type='{__get_event_type(event_type, platform=platform)}'")
|
||||||
|
events_conditions.append({"type": event_where[-1]})
|
||||||
|
if not is_any:
|
||||||
|
if is_not:
|
||||||
|
event_where.append(_multiple_conditions(f"sub.{_column} {op} %({e_k})s", event.value,
|
||||||
|
value_key=e_k))
|
||||||
|
events_conditions_not.append(
|
||||||
|
{"type": f"sub.event_type='{__get_event_type(event_type, platform=platform)}'"})
|
||||||
|
events_conditions_not[-1]["condition"] = event_where[-1]
|
||||||
|
else:
|
||||||
|
event_where.append(_multiple_conditions(f"main.{_column} {op} %({e_k})s", event.value,
|
||||||
|
value_key=e_k))
|
||||||
|
events_conditions[-1]["condition"] = event_where[-1]
|
||||||
|
elif event_type == events.EventType.INPUT_MOBILE.ui_type:
|
||||||
|
_column = events.EventType.INPUT_MOBILE.column
|
||||||
|
event_where.append(f"main.event_type='{__get_event_type(event_type, platform=platform)}'")
|
||||||
|
events_conditions.append({"type": event_where[-1]})
|
||||||
|
if not is_any:
|
||||||
|
if is_not:
|
||||||
|
event_where.append(_multiple_conditions(f"sub.{_column} {op} %({e_k})s", event.value,
|
||||||
|
value_key=e_k))
|
||||||
|
events_conditions_not.append(
|
||||||
|
{"type": f"sub.event_type='{__get_event_type(event_type, platform=platform)}'"})
|
||||||
|
events_conditions_not[-1]["condition"] = event_where[-1]
|
||||||
|
else:
|
||||||
|
event_where.append(_multiple_conditions(f"main.{_column} {op} %({e_k})s", event.value,
|
||||||
|
value_key=e_k))
|
||||||
|
events_conditions[-1]["condition"] = event_where[-1]
|
||||||
|
elif event_type == events.EventType.VIEW_MOBILE.ui_type:
|
||||||
|
_column = events.EventType.VIEW_MOBILE.column
|
||||||
|
event_where.append(f"main.event_type='{__get_event_type(event_type, platform=platform)}'")
|
||||||
|
events_conditions.append({"type": event_where[-1]})
|
||||||
|
if not is_any:
|
||||||
|
if is_not:
|
||||||
|
event_where.append(_multiple_conditions(f"sub.{_column} {op} %({e_k})s", event.value,
|
||||||
|
value_key=e_k))
|
||||||
|
events_conditions_not.append(
|
||||||
|
{"type": f"sub.event_type='{__get_event_type(event_type, platform=platform)}'"})
|
||||||
|
events_conditions_not[-1]["condition"] = event_where[-1]
|
||||||
|
else:
|
||||||
|
event_where.append(_multiple_conditions(f"main.{_column} {op} %({e_k})s",
|
||||||
|
event.value, value_key=e_k))
|
||||||
|
events_conditions[-1]["condition"] = event_where[-1]
|
||||||
|
elif event_type == events.EventType.CUSTOM_MOBILE.ui_type:
|
||||||
|
_column = events.EventType.CUSTOM_MOBILE.column
|
||||||
|
event_where.append(f"main.event_type='{__get_event_type(event_type, platform=platform)}'")
|
||||||
|
events_conditions.append({"type": event_where[-1]})
|
||||||
|
if not is_any:
|
||||||
|
if is_not:
|
||||||
|
event_where.append(_multiple_conditions(f"sub.{_column} {op} %({e_k})s", event.value,
|
||||||
|
value_key=e_k))
|
||||||
|
events_conditions_not.append(
|
||||||
|
{"type": f"sub.event_type='{__get_event_type(event_type, platform=platform)}'"})
|
||||||
|
events_conditions_not[-1]["condition"] = event_where[-1]
|
||||||
|
else:
|
||||||
|
event_where.append(_multiple_conditions(f"main.{_column} {op} %({e_k})s",
|
||||||
|
event.value, value_key=e_k))
|
||||||
|
events_conditions[-1]["condition"] = event_where[-1]
|
||||||
|
elif event_type == events.EventType.REQUEST_MOBILE.ui_type:
|
||||||
|
event_from = event_from % f"{MAIN_EVENTS_TABLE} AS main "
|
||||||
|
_column = 'url_path'
|
||||||
|
event_where.append(f"main.event_type='{__get_event_type(event_type, platform=platform)}'")
|
||||||
|
events_conditions.append({"type": event_where[-1]})
|
||||||
|
if not is_any:
|
||||||
|
if is_not:
|
||||||
|
event_where.append(_multiple_conditions(f"sub.{_column} {op} %({e_k})s", event.value,
|
||||||
|
value_key=e_k))
|
||||||
|
events_conditions_not.append(
|
||||||
|
{"type": f"sub.event_type='{__get_event_type(event_type, platform=platform)}'"})
|
||||||
|
events_conditions_not[-1]["condition"] = event_where[-1]
|
||||||
|
else:
|
||||||
|
event_where.append(_multiple_conditions(f"main.{_column} {op} %({e_k})s", event.value,
|
||||||
|
value_key=e_k))
|
||||||
|
events_conditions[-1]["condition"] = event_where[-1]
|
||||||
|
elif event_type == events.EventType.CRASH_MOBILE.ui_type:
|
||||||
|
_column = events.EventType.CRASH_MOBILE.column
|
||||||
|
event_where.append(f"main.event_type='{__get_event_type(event_type, platform=platform)}'")
|
||||||
|
events_conditions.append({"type": event_where[-1]})
|
||||||
|
if not is_any:
|
||||||
|
if is_not:
|
||||||
|
event_where.append(_multiple_conditions(f"sub.{_column} {op} %({e_k})s", event.value,
|
||||||
|
value_key=e_k))
|
||||||
|
events_conditions_not.append(
|
||||||
|
{"type": f"sub.event_type='{__get_event_type(event_type, platform=platform)}'"})
|
||||||
|
events_conditions_not[-1]["condition"] = event_where[-1]
|
||||||
|
else:
|
||||||
|
event_where.append(_multiple_conditions(f"main.{_column} {op} %({e_k})s",
|
||||||
|
event.value, value_key=e_k))
|
||||||
|
events_conditions[-1]["condition"] = event_where[-1]
|
||||||
|
elif event_type == events.EventType.SWIPE_MOBILE.ui_type and platform != "web":
|
||||||
|
_column = events.EventType.SWIPE_MOBILE.column
|
||||||
|
event_where.append(f"main.event_type='{__get_event_type(event_type, platform=platform)}'")
|
||||||
|
events_conditions.append({"type": event_where[-1]})
|
||||||
|
if not is_any:
|
||||||
|
if is_not:
|
||||||
|
event_where.append(_multiple_conditions(f"sub.{_column} {op} %({e_k})s", event.value,
|
||||||
|
value_key=e_k))
|
||||||
|
events_conditions_not.append(
|
||||||
|
{"type": f"sub.event_type='{__get_event_type(event_type, platform=platform)}'"})
|
||||||
|
events_conditions_not[-1]["condition"] = event_where[-1]
|
||||||
|
else:
|
||||||
|
event_where.append(_multiple_conditions(f"main.{_column} {op} %({e_k})s",
|
||||||
|
event.value, value_key=e_k))
|
||||||
|
events_conditions[-1]["condition"] = event_where[-1]
|
||||||
|
|
||||||
elif event_type == schemas.PerformanceEventType.fetch_failed:
|
elif event_type == schemas.PerformanceEventType.fetch_failed:
|
||||||
event_from = event_from % f"{MAIN_EVENTS_TABLE} AS main "
|
event_from = event_from % f"{MAIN_EVENTS_TABLE} AS main "
|
||||||
_column = 'url_path'
|
_column = 'url_path'
|
||||||
|
|
|
||||||
|
|
@ -185,8 +185,9 @@ def __filter_subquery(project_id: int, filters: Optional[schemas.SessionsSearchP
|
||||||
errors_only=True, favorite_only=None,
|
errors_only=True, favorite_only=None,
|
||||||
issue=None, user_id=None)
|
issue=None, user_id=None)
|
||||||
params = {**params, **qp_params}
|
params = {**params, **qp_params}
|
||||||
# TODO: test if this line impacts other cards beside insights
|
# This line was added because insights is failing when you add filter steps,
|
||||||
# sub_query = f"INNER JOIN {sub_query} USING(session_id)"
|
# for example when you add a LOCATION filter
|
||||||
|
sub_query = f"INNER JOIN {sub_query} USING(session_id)"
|
||||||
return params, sub_query
|
return params, sub_query
|
||||||
|
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -397,12 +397,19 @@ func (c *connectorImpl) InsertWebClickEvent(session *sessions.Session, msg *mess
|
||||||
if msg.Label == "" {
|
if msg.Label == "" {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
var nX *uint8 = nil
|
var nX *float32 = nil
|
||||||
var nY *uint8 = nil
|
var nY *float32 = nil
|
||||||
if msg.NormalizedX <= 100 && msg.NormalizedY <= 100 {
|
if msg.NormalizedX != 101 && msg.NormalizedY != 101 {
|
||||||
nXVal := uint8(msg.NormalizedX)
|
// To support previous versions of tracker
|
||||||
|
if msg.NormalizedX <= 100 && msg.NormalizedY <= 100 {
|
||||||
|
msg.NormalizedX *= 100
|
||||||
|
msg.NormalizedY *= 100
|
||||||
|
}
|
||||||
|
normalizedX := float32(msg.NormalizedX) / 100.0
|
||||||
|
normalizedY := float32(msg.NormalizedY) / 100.0
|
||||||
|
nXVal := normalizedX
|
||||||
nX = &nXVal
|
nX = &nXVal
|
||||||
nYVal := uint8(msg.NormalizedY)
|
nYVal := normalizedY
|
||||||
nY = &nYVal
|
nY = &nYVal
|
||||||
}
|
}
|
||||||
if err := c.batches["clicks"].Append(
|
if err := c.batches["clicks"].Append(
|
||||||
|
|
|
||||||
|
|
@ -3,8 +3,8 @@ CREATE OR REPLACE FUNCTION openreplay_version AS() -> 'v1.19.0-ee';
|
||||||
DROP TABLE IF EXISTS experimental.events_l7d_mv;
|
DROP TABLE IF EXISTS experimental.events_l7d_mv;
|
||||||
|
|
||||||
ALTER TABLE experimental.events
|
ALTER TABLE experimental.events
|
||||||
ADD COLUMN IF NOT EXISTS normalized_x Nullable(UInt8),
|
ADD COLUMN IF NOT EXISTS normalized_x Nullable(Float32),
|
||||||
ADD COLUMN IF NOT EXISTS normalized_y Nullable(UInt8),
|
ADD COLUMN IF NOT EXISTS normalized_y Nullable(Float32),
|
||||||
DROP COLUMN IF EXISTS coordinate;
|
DROP COLUMN IF EXISTS coordinate;
|
||||||
|
|
||||||
CREATE MATERIALIZED VIEW IF NOT EXISTS experimental.events_l7d_mv
|
CREATE MATERIALIZED VIEW IF NOT EXISTS experimental.events_l7d_mv
|
||||||
|
|
|
||||||
|
|
@ -81,8 +81,8 @@ CREATE TABLE IF NOT EXISTS experimental.events
|
||||||
error_tags_values Array(Nullable(String)),
|
error_tags_values Array(Nullable(String)),
|
||||||
transfer_size Nullable(UInt32),
|
transfer_size Nullable(UInt32),
|
||||||
selector Nullable(String),
|
selector Nullable(String),
|
||||||
normalized_x Nullable(UInt8),
|
normalized_x Nullable(Float32),
|
||||||
normalized_y Nullable(UInt8),
|
normalized_y Nullable(Float32),
|
||||||
message_id UInt64 DEFAULT 0,
|
message_id UInt64 DEFAULT 0,
|
||||||
_timestamp DateTime DEFAULT now()
|
_timestamp DateTime DEFAULT now()
|
||||||
) ENGINE = ReplacingMergeTree(_timestamp)
|
) ENGINE = ReplacingMergeTree(_timestamp)
|
||||||
|
|
|
||||||
|
|
@ -19,8 +19,8 @@ $fn_def$, :'next_version')
|
||||||
|
|
||||||
--
|
--
|
||||||
ALTER TABLE IF EXISTS events.clicks
|
ALTER TABLE IF EXISTS events.clicks
|
||||||
ADD COLUMN IF NOT EXISTS normalized_x smallint NULL,
|
ADD COLUMN IF NOT EXISTS normalized_x decimal NULL,
|
||||||
ADD COLUMN IF NOT EXISTS normalized_y smallint NULL,
|
ADD COLUMN IF NOT EXISTS normalized_y decimal NULL,
|
||||||
DROP COLUMN IF EXISTS x,
|
DROP COLUMN IF EXISTS x,
|
||||||
DROP COLUMN IF EXISTS y;
|
DROP COLUMN IF EXISTS y;
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -659,16 +659,16 @@ CREATE INDEX pages_query_nn_gin_idx ON events.pages USING GIN (query gin_trgm_op
|
||||||
|
|
||||||
CREATE TABLE events.clicks
|
CREATE TABLE events.clicks
|
||||||
(
|
(
|
||||||
session_id bigint NOT NULL REFERENCES public.sessions (session_id) ON DELETE CASCADE,
|
session_id bigint NOT NULL REFERENCES public.sessions (session_id) ON DELETE CASCADE,
|
||||||
message_id bigint NOT NULL,
|
message_id bigint NOT NULL,
|
||||||
timestamp bigint NOT NULL,
|
timestamp bigint NOT NULL,
|
||||||
label text DEFAULT NULL,
|
label text DEFAULT NULL,
|
||||||
url text DEFAULT '' NOT NULL,
|
url text DEFAULT '' NOT NULL,
|
||||||
path text,
|
path text,
|
||||||
selector text DEFAULT '' NOT NULL,
|
selector text DEFAULT '' NOT NULL,
|
||||||
hesitation integer DEFAULT NULL,
|
hesitation integer DEFAULT NULL,
|
||||||
normalized_x smallint DEFAULT NULL,
|
normalized_x decimal DEFAULT NULL,
|
||||||
normalized_y smallint DEFAULT NULL,
|
normalized_y decimal DEFAULT NULL,
|
||||||
PRIMARY KEY (session_id, message_id)
|
PRIMARY KEY (session_id, message_id)
|
||||||
);
|
);
|
||||||
CREATE INDEX clicks_session_id_idx ON events.clicks (session_id);
|
CREATE INDEX clicks_session_id_idx ON events.clicks (session_id);
|
||||||
|
|
|
||||||
|
|
@ -62,9 +62,9 @@ function ClickMapCard({
|
||||||
if (mapUrl) return evt.path.includes(mapUrl)
|
if (mapUrl) return evt.path.includes(mapUrl)
|
||||||
return evt
|
return evt
|
||||||
}) || { timestamp: metricStore.instance.data.startTs }
|
}) || { timestamp: metricStore.instance.data.startTs }
|
||||||
|
const ts = jumpToEvent.timestamp ?? metricStore.instance.data.startTs
|
||||||
const jumpTimestamp = (jumpToEvent.timestamp - metricStore.instance.data.startTs) + jumpToEvent.domBuildingTime + 99 // 99ms safety margin to give some time for the DOM to load
|
const domTime = jumpToEvent.domBuildingTime ?? 0
|
||||||
|
const jumpTimestamp = (ts - metricStore.instance.data.startTs) + domTime + 99 // 99ms safety margin to give some time for the DOM to load
|
||||||
return (
|
return (
|
||||||
<div id="clickmap-render">
|
<div id="clickmap-render">
|
||||||
<ClickMapRenderer
|
<ClickMapRenderer
|
||||||
|
|
|
||||||
|
|
@ -51,8 +51,8 @@ function WebPlayer(props: any) {
|
||||||
const isPlayerReady = contextValue.store?.get().ready
|
const isPlayerReady = contextValue.store?.get().ready
|
||||||
|
|
||||||
React.useEffect(() => {
|
React.useEffect(() => {
|
||||||
contextValue.player && contextValue.player.play()
|
contextValue.player && contextValue.player.play()
|
||||||
if (isPlayerReady && insights.size > 0) {
|
if (isPlayerReady && insights.size > 0 && jumpTimestamp) {
|
||||||
setTimeout(() => {
|
setTimeout(() => {
|
||||||
contextValue.player.pause()
|
contextValue.player.pause()
|
||||||
contextValue.player.jump(jumpTimestamp)
|
contextValue.player.jump(jumpTimestamp)
|
||||||
|
|
|
||||||
|
|
@ -9,7 +9,7 @@
|
||||||
height: 100%;
|
height: 100%;
|
||||||
/* border: solid thin $gray-light; */
|
/* border: solid thin $gray-light; */
|
||||||
/* border-radius: 3px; */
|
/* border-radius: 3px; */
|
||||||
overflow: hidden;
|
overflow-y: scroll;
|
||||||
}
|
}
|
||||||
|
|
||||||
.checkers {
|
.checkers {
|
||||||
|
|
|
||||||
|
|
@ -302,8 +302,8 @@ export default class Widget {
|
||||||
} else if (this.metricType === FUNNEL) {
|
} else if (this.metricType === FUNNEL) {
|
||||||
_data.funnel = new Funnel().fromJSON(_data);
|
_data.funnel = new Funnel().fromJSON(_data);
|
||||||
} else if (this.metricType === TABLE) {
|
} else if (this.metricType === TABLE) {
|
||||||
const totalSessions = data[0]['totalSessions'];
|
const count = data[0]['count'];
|
||||||
_data[0]['values'] = data[0]['values'].map((s: any) => new SessionsByRow().fromJson(s, totalSessions, this.metricOf));
|
_data[0]['values'] = data[0]['values'].map((s: any) => new SessionsByRow().fromJson(s, count, this.metricOf));
|
||||||
} else {
|
} else {
|
||||||
if (data.hasOwnProperty('chart')) {
|
if (data.hasOwnProperty('chart')) {
|
||||||
_data['value'] = data.value;
|
_data['value'] = data.value;
|
||||||
|
|
|
||||||
|
|
@ -78,10 +78,10 @@ export interface State extends ScreenState, ListsState {
|
||||||
}
|
}
|
||||||
|
|
||||||
const userEvents = [
|
const userEvents = [
|
||||||
MType.IosSwipeEvent,
|
MType.MobileSwipeEvent,
|
||||||
MType.IosClickEvent,
|
MType.MobileClickEvent,
|
||||||
MType.IosInputEvent,
|
MType.MobileInputEvent,
|
||||||
MType.IosScreenChanges,
|
MType.MobileScreenChanges,
|
||||||
];
|
];
|
||||||
|
|
||||||
export default class IOSMessageManager implements IMessageManager {
|
export default class IOSMessageManager implements IMessageManager {
|
||||||
|
|
@ -233,7 +233,7 @@ export default class IOSMessageManager implements IMessageManager {
|
||||||
}
|
}
|
||||||
|
|
||||||
switch (msg.tp) {
|
switch (msg.tp) {
|
||||||
case MType.IosPerformanceEvent:
|
case MType.MobilePerformanceEvent:
|
||||||
const performanceStats = ['background', 'memoryUsage', 'mainThreadCPU'];
|
const performanceStats = ['background', 'memoryUsage', 'mainThreadCPU'];
|
||||||
if (performanceStats.includes(msg.name)) {
|
if (performanceStats.includes(msg.name)) {
|
||||||
this.performanceManager.append(msg);
|
this.performanceManager.append(msg);
|
||||||
|
|
@ -253,21 +253,21 @@ export default class IOSMessageManager implements IMessageManager {
|
||||||
// case MType.IosInputEvent:
|
// case MType.IosInputEvent:
|
||||||
// console.log('input', msg)
|
// console.log('input', msg)
|
||||||
// break;
|
// break;
|
||||||
case MType.IosNetworkCall:
|
case MType.MobileNetworkCall:
|
||||||
this.lists.lists.fetch.insert(getResourceFromNetworkRequest(msg, this.sessionStart));
|
this.lists.lists.fetch.insert(getResourceFromNetworkRequest(msg, this.sessionStart));
|
||||||
break;
|
break;
|
||||||
case MType.WsChannel:
|
case MType.WsChannel:
|
||||||
this.lists.lists.websocket.insert(msg);
|
this.lists.lists.websocket.insert(msg);
|
||||||
break;
|
break;
|
||||||
case MType.IosEvent:
|
case MType.MobileEvent:
|
||||||
// @ts-ignore
|
// @ts-ignore
|
||||||
this.lists.lists.event.insert({ ...msg, source: 'openreplay' });
|
this.lists.lists.event.insert({ ...msg, source: 'openreplay' });
|
||||||
break;
|
break;
|
||||||
case MType.IosSwipeEvent:
|
case MType.MobileSwipeEvent:
|
||||||
case MType.IosClickEvent:
|
case MType.MobileClickEvent:
|
||||||
this.touchManager.append(msg);
|
this.touchManager.append(msg);
|
||||||
break;
|
break;
|
||||||
case MType.IosLog:
|
case MType.MobileLog:
|
||||||
const log = { ...msg, level: msg.severity };
|
const log = { ...msg, level: msg.severity };
|
||||||
// @ts-ignore
|
// @ts-ignore
|
||||||
this.lists.lists.log.append(Log(log));
|
this.lists.lists.log.append(Log(log));
|
||||||
|
|
|
||||||
|
|
@ -31,7 +31,7 @@ export default class TouchManager extends ListWalker<IosClickEvent | IosSwipeEve
|
||||||
public move(t: number) {
|
public move(t: number) {
|
||||||
const lastTouch = this.moveGetLast(t)
|
const lastTouch = this.moveGetLast(t)
|
||||||
if (!!lastTouch) {
|
if (!!lastTouch) {
|
||||||
if (lastTouch.tp === MType.IosSwipeEvent) {
|
if (lastTouch.tp === MType.MobileSwipeEvent) {
|
||||||
return
|
return
|
||||||
// not using swipe rn
|
// not using swipe rn
|
||||||
// this.touchTrail?.createSwipeTrail({
|
// this.touchTrail?.createSwipeTrail({
|
||||||
|
|
|
||||||
|
|
@ -233,10 +233,10 @@ export default class Screen {
|
||||||
break;
|
break;
|
||||||
case ScaleMode.AdjustParentHeight:
|
case ScaleMode.AdjustParentHeight:
|
||||||
// we want to scale the document with true height so the clickmap will be scrollable
|
// we want to scale the document with true height so the clickmap will be scrollable
|
||||||
const usedHeight =
|
const usedHeight = height + 'px';
|
||||||
this.document?.body.scrollHeight && this.document?.body.scrollHeight > height
|
// this.document?.body.scrollHeight && this.document?.body.scrollHeight > height
|
||||||
? this.document.body.scrollHeight + 'px'
|
// ? this.document.body.scrollHeight + 'px'
|
||||||
: height + 'px';
|
// : height + 'px';
|
||||||
this.scaleRatio = offsetWidth / width;
|
this.scaleRatio = offsetWidth / width;
|
||||||
translate = 'translate(-50%, 0)';
|
translate = 'translate(-50%, 0)';
|
||||||
posStyles = { top: 0, height: usedHeight };
|
posStyles = { top: 0, height: usedHeight };
|
||||||
|
|
|
||||||
|
|
@ -146,39 +146,37 @@ export default class TargetMarker {
|
||||||
if (clicks && this.screen.document) {
|
if (clicks && this.screen.document) {
|
||||||
this.clickMapOverlay?.remove();
|
this.clickMapOverlay?.remove();
|
||||||
const overlay = document.createElement('canvas');
|
const overlay = document.createElement('canvas');
|
||||||
const iframeSize = this.screen.iframeStylesRef;
|
|
||||||
const scrollHeight = this.screen.document?.documentElement.scrollHeight || 0;
|
const scrollHeight = this.screen.document?.documentElement.scrollHeight || 0;
|
||||||
const scrollWidth = this.screen.document?.documentElement.scrollWidth || 0;
|
const scrollWidth = this.screen.document?.documentElement.scrollWidth || 0;
|
||||||
const scaleRatio = this.screen.getScale();
|
|
||||||
Object.assign(
|
Object.assign(
|
||||||
overlay.style,
|
overlay.style,
|
||||||
clickmapStyles.overlayStyle({
|
clickmapStyles.overlayStyle({
|
||||||
height: iframeSize.height,
|
height: scrollHeight + 'px',
|
||||||
width: iframeSize.width,
|
width: scrollWidth + 'px',
|
||||||
scale: scaleRatio,
|
|
||||||
})
|
})
|
||||||
);
|
);
|
||||||
|
|
||||||
this.clickMapOverlay = overlay;
|
this.clickMapOverlay = overlay;
|
||||||
this.screen.getParentElement()?.appendChild(overlay);
|
this.screen.document.body.appendChild(overlay);
|
||||||
|
|
||||||
const pointMap: Record<string, { times: number; data: number[], original: any }> = {};
|
const pointMap: Record<string, { times: number; data: number[], original: any }> = {};
|
||||||
const ovWidth = parseInt(iframeSize.width);
|
overlay.width = scrollWidth;
|
||||||
const ovHeight = parseInt(iframeSize.height);
|
overlay.height = scrollHeight;
|
||||||
overlay.width = ovWidth;
|
|
||||||
overlay.height = ovHeight;
|
|
||||||
let maxIntensity = 0;
|
let maxIntensity = 0;
|
||||||
|
|
||||||
clicks.forEach((point) => {
|
clicks.forEach((point) => {
|
||||||
const key = `${point.normalizedY}-${point.normalizedX}`;
|
const y = roundToSecond(point.normalizedY);
|
||||||
|
const x = roundToSecond(point.normalizedX);
|
||||||
|
const key = `${y}-${x}`;
|
||||||
if (pointMap[key]) {
|
if (pointMap[key]) {
|
||||||
const times = pointMap[key].times + 1;
|
const times = pointMap[key].times + 1;
|
||||||
maxIntensity = Math.max(maxIntensity, times);
|
maxIntensity = Math.max(maxIntensity, times);
|
||||||
pointMap[key].times = times;
|
pointMap[key].times = times;
|
||||||
} else {
|
} else {
|
||||||
const clickData = [
|
const clickData = [
|
||||||
(point.normalizedX / 100) * scrollWidth,
|
(x / 100) * scrollWidth,
|
||||||
(point.normalizedY / 100) * scrollHeight,
|
(y / 100) * scrollHeight,
|
||||||
];
|
];
|
||||||
pointMap[key] = { times: 1, data: clickData, original: point };
|
pointMap[key] = { times: 1, data: clickData, original: point };
|
||||||
}
|
}
|
||||||
|
|
@ -204,3 +202,7 @@ export default class TargetMarker {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
function roundToSecond(num: number) {
|
||||||
|
return Math.round(num * 100) / 100;
|
||||||
|
}
|
||||||
|
|
@ -1,14 +1,12 @@
|
||||||
export const clickmapStyles = {
|
export const clickmapStyles = {
|
||||||
overlayStyle: ({ height, width, scale }: { height: string, width: string, scale: number }) => ({
|
overlayStyle: ({ height, width }: { height: string, width: string }) => ({
|
||||||
transform: `scale(${scale}) translate(-50%, 0)`,
|
|
||||||
position: 'absolute',
|
position: 'absolute',
|
||||||
top: '0px',
|
top: '0px',
|
||||||
left: '50%',
|
left: 0,
|
||||||
width,
|
width,
|
||||||
height,
|
height,
|
||||||
background: 'rgba(0,0,0, 0.15)',
|
background: 'rgba(0,0,0, 0.15)',
|
||||||
zIndex: 9 * 10e3,
|
zIndex: 9 * 10e3,
|
||||||
transformOrigin: 'left top',
|
|
||||||
}),
|
}),
|
||||||
totalClicks: {
|
totalClicks: {
|
||||||
fontSize: '16px',
|
fontSize: '16px',
|
||||||
|
|
|
||||||
|
|
@ -1,7 +1,6 @@
|
||||||
apiVersion: v2
|
apiVersion: v2
|
||||||
name: chalice
|
name: chalice
|
||||||
description: A Helm chart for Kubernetes
|
description: A Helm chart for Kubernetes
|
||||||
|
|
||||||
# A chart can be either an 'application' or a 'library' chart.
|
# A chart can be either an 'application' or a 'library' chart.
|
||||||
#
|
#
|
||||||
# Application charts are a collection of templates that can be packaged into versioned archives
|
# Application charts are a collection of templates that can be packaged into versioned archives
|
||||||
|
|
@ -11,14 +10,12 @@ description: A Helm chart for Kubernetes
|
||||||
# a dependency of application charts to inject those utilities and functions into the rendering
|
# a dependency of application charts to inject those utilities and functions into the rendering
|
||||||
# pipeline. Library charts do not define any templates and therefore cannot be deployed.
|
# pipeline. Library charts do not define any templates and therefore cannot be deployed.
|
||||||
type: application
|
type: application
|
||||||
|
|
||||||
# This is the chart version. This version number should be incremented each time you make changes
|
# This is the chart version. This version number should be incremented each time you make changes
|
||||||
# to the chart and its templates, including the app version.
|
# to the chart and its templates, including the app version.
|
||||||
# Versions are expected to follow Semantic Versioning (https://semver.org/)
|
# Versions are expected to follow Semantic Versioning (https://semver.org/)
|
||||||
version: 0.1.7
|
version: 0.1.7
|
||||||
|
|
||||||
# This is the version number of the application being deployed. This version number should be
|
# This is the version number of the application being deployed. This version number should be
|
||||||
# incremented each time you make changes to the application. Versions are not expected to
|
# incremented each time you make changes to the application. Versions are not expected to
|
||||||
# follow Semantic Versioning. They should reflect the version the application is using.
|
# follow Semantic Versioning. They should reflect the version the application is using.
|
||||||
# It is recommended to use it with quotes.
|
# It is recommended to use it with quotes.
|
||||||
AppVersion: "v1.19.0"
|
AppVersion: "v1.19.6"
|
||||||
|
|
|
||||||
|
|
@ -1,7 +1,6 @@
|
||||||
apiVersion: v2
|
apiVersion: v2
|
||||||
name: db
|
name: db
|
||||||
description: A Helm chart for Kubernetes
|
description: A Helm chart for Kubernetes
|
||||||
|
|
||||||
# A chart can be either an 'application' or a 'library' chart.
|
# A chart can be either an 'application' or a 'library' chart.
|
||||||
#
|
#
|
||||||
# Application charts are a collection of templates that can be packaged into versioned archives
|
# Application charts are a collection of templates that can be packaged into versioned archives
|
||||||
|
|
@ -11,14 +10,12 @@ description: A Helm chart for Kubernetes
|
||||||
# a dependency of application charts to inject those utilities and functions into the rendering
|
# a dependency of application charts to inject those utilities and functions into the rendering
|
||||||
# pipeline. Library charts do not define any templates and therefore cannot be deployed.
|
# pipeline. Library charts do not define any templates and therefore cannot be deployed.
|
||||||
type: application
|
type: application
|
||||||
|
|
||||||
# This is the chart version. This version number should be incremented each time you make changes
|
# This is the chart version. This version number should be incremented each time you make changes
|
||||||
# to the chart and its templates, including the app version.
|
# to the chart and its templates, including the app version.
|
||||||
# Versions are expected to follow Semantic Versioning (https://semver.org/)
|
# Versions are expected to follow Semantic Versioning (https://semver.org/)
|
||||||
version: 0.1.1
|
version: 0.1.1
|
||||||
|
|
||||||
# This is the version number of the application being deployed. This version number should be
|
# This is the version number of the application being deployed. This version number should be
|
||||||
# incremented each time you make changes to the application. Versions are not expected to
|
# incremented each time you make changes to the application. Versions are not expected to
|
||||||
# follow Semantic Versioning. They should reflect the version the application is using.
|
# follow Semantic Versioning. They should reflect the version the application is using.
|
||||||
# It is recommended to use it with quotes.
|
# It is recommended to use it with quotes.
|
||||||
AppVersion: "v1.19.0"
|
AppVersion: "v1.19.1"
|
||||||
|
|
|
||||||
|
|
@ -1,7 +1,6 @@
|
||||||
apiVersion: v2
|
apiVersion: v2
|
||||||
name: frontend
|
name: frontend
|
||||||
description: A Helm chart for Kubernetes
|
description: A Helm chart for Kubernetes
|
||||||
|
|
||||||
# A chart can be either an 'application' or a 'library' chart.
|
# A chart can be either an 'application' or a 'library' chart.
|
||||||
#
|
#
|
||||||
# Application charts are a collection of templates that can be packaged into versioned archives
|
# Application charts are a collection of templates that can be packaged into versioned archives
|
||||||
|
|
@ -11,14 +10,12 @@ description: A Helm chart for Kubernetes
|
||||||
# a dependency of application charts to inject those utilities and functions into the rendering
|
# a dependency of application charts to inject those utilities and functions into the rendering
|
||||||
# pipeline. Library charts do not define any templates and therefore cannot be deployed.
|
# pipeline. Library charts do not define any templates and therefore cannot be deployed.
|
||||||
type: application
|
type: application
|
||||||
|
|
||||||
# This is the chart version. This version number should be incremented each time you make changes
|
# This is the chart version. This version number should be incremented each time you make changes
|
||||||
# to the chart and its templates, including the app version.
|
# to the chart and its templates, including the app version.
|
||||||
# Versions are expected to follow Semantic Versioning (frontends://semver.org/)
|
# Versions are expected to follow Semantic Versioning (frontends://semver.org/)
|
||||||
version: 0.1.10
|
version: 0.1.10
|
||||||
|
|
||||||
# This is the version number of the application being deployed. This version number should be
|
# This is the version number of the application being deployed. This version number should be
|
||||||
# incremented each time you make changes to the application. Versions are not expected to
|
# incremented each time you make changes to the application. Versions are not expected to
|
||||||
# follow Semantic Versioning. They should reflect the version the application is using.
|
# follow Semantic Versioning. They should reflect the version the application is using.
|
||||||
# It is recommended to use it with quotes.
|
# It is recommended to use it with quotes.
|
||||||
AppVersion: "v1.19.0"
|
AppVersion: "v1.19.2"
|
||||||
|
|
|
||||||
|
|
@ -8,7 +8,11 @@ RETENTION_TIME=${RETENTION_TIME:-345600000}
|
||||||
topics=(
|
topics=(
|
||||||
"raw"
|
"raw"
|
||||||
"raw-ios"
|
"raw-ios"
|
||||||
|
"raw-images"
|
||||||
|
"canvas-images"
|
||||||
"trigger"
|
"trigger"
|
||||||
|
"canvas-trigger"
|
||||||
|
"mobile-trigger"
|
||||||
"cache"
|
"cache"
|
||||||
"analytics"
|
"analytics"
|
||||||
"storage-failover"
|
"storage-failover"
|
||||||
|
|
|
||||||
|
|
@ -19,8 +19,8 @@ $fn_def$, :'next_version')
|
||||||
|
|
||||||
--
|
--
|
||||||
ALTER TABLE IF EXISTS events.clicks
|
ALTER TABLE IF EXISTS events.clicks
|
||||||
ADD COLUMN IF NOT EXISTS normalized_x smallint NULL,
|
ADD COLUMN IF NOT EXISTS normalized_x decimal NULL,
|
||||||
ADD COLUMN IF NOT EXISTS normalized_y smallint NULL,
|
ADD COLUMN IF NOT EXISTS normalized_y decimal NULL,
|
||||||
DROP COLUMN IF EXISTS x,
|
DROP COLUMN IF EXISTS x,
|
||||||
DROP COLUMN IF EXISTS y;
|
DROP COLUMN IF EXISTS y;
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -620,16 +620,16 @@ CREATE INDEX pages_query_nn_gin_idx ON events.pages USING GIN (query gin_trgm_op
|
||||||
|
|
||||||
CREATE TABLE events.clicks
|
CREATE TABLE events.clicks
|
||||||
(
|
(
|
||||||
session_id bigint NOT NULL REFERENCES public.sessions (session_id) ON DELETE CASCADE,
|
session_id bigint NOT NULL REFERENCES public.sessions (session_id) ON DELETE CASCADE,
|
||||||
message_id bigint NOT NULL,
|
message_id bigint NOT NULL,
|
||||||
timestamp bigint NOT NULL,
|
timestamp bigint NOT NULL,
|
||||||
label text DEFAULT NULL,
|
label text DEFAULT NULL,
|
||||||
url text DEFAULT '' NOT NULL,
|
url text DEFAULT '' NOT NULL,
|
||||||
path text,
|
path text,
|
||||||
selector text DEFAULT '' NOT NULL,
|
selector text DEFAULT '' NOT NULL,
|
||||||
hesitation integer DEFAULT NULL,
|
hesitation integer DEFAULT NULL,
|
||||||
normalized_x smallint DEFAULT NULL,
|
normalized_x decimal DEFAULT NULL,
|
||||||
normalized_y smallint DEFAULT NULL,
|
normalized_y decimal DEFAULT NULL,
|
||||||
PRIMARY KEY (session_id, message_id)
|
PRIMARY KEY (session_id, message_id)
|
||||||
);
|
);
|
||||||
CREATE INDEX clicks_session_id_idx ON events.clicks (session_id);
|
CREATE INDEX clicks_session_id_idx ON events.clicks (session_id);
|
||||||
|
|
|
||||||
|
|
@ -18,13 +18,13 @@ returns `result` without changes.
|
||||||
|
|
||||||
```js
|
```js
|
||||||
import Tracker from '@openreplay/tracker';
|
import Tracker from '@openreplay/tracker';
|
||||||
import trackerGraphQL from '@openreplay/tracker-graphql';
|
import { createGraphqlMiddleware } from '@openreplay/tracker-graphql';
|
||||||
|
|
||||||
const tracker = new Tracker({
|
const tracker = new Tracker({
|
||||||
projectKey: YOUR_PROJECT_KEY,
|
projectKey: YOUR_PROJECT_KEY,
|
||||||
});
|
});
|
||||||
|
|
||||||
export const recordGraphQL = tracker.plugin(trackerGraphQL());
|
export const recordGraphQL = tracker.use(createGraphqlMiddleware());
|
||||||
```
|
```
|
||||||
|
|
||||||
### Relay
|
### Relay
|
||||||
|
|
@ -33,15 +33,28 @@ If you're using [Relay network tools](https://github.com/relay-tools/react-relay
|
||||||
you can simply [create a middleware](https://github.com/relay-tools/react-relay-network-modern/tree/master?tab=readme-ov-file#example-of-injecting-networklayer-with-middlewares-on-the-client-side)
|
you can simply [create a middleware](https://github.com/relay-tools/react-relay-network-modern/tree/master?tab=readme-ov-file#example-of-injecting-networklayer-with-middlewares-on-the-client-side)
|
||||||
|
|
||||||
```js
|
```js
|
||||||
import { createRelayMiddleware } from '@openreplay/tracker-graphql'
|
import { createRelayMiddleware } from '@openreplay/tracker-graphql';
|
||||||
|
|
||||||
const trackerMiddleware = createRelayMiddleware(tracker)
|
const trackerMiddleware = tracker.use(createRelayMiddleware());
|
||||||
|
|
||||||
const network = new RelayNetworkLayer([
|
const network = new RelayNetworkLayer([
|
||||||
// your middleware
|
// your middleware
|
||||||
// ,
|
// ,
|
||||||
trackerMiddleware
|
trackerMiddleware,
|
||||||
])
|
]);
|
||||||
|
```
|
||||||
|
|
||||||
|
You can pass a Sanitizer function to `createRelayMiddleware` to sanitize the variables and data before sending them to OpenReplay.
|
||||||
|
|
||||||
|
```js
|
||||||
|
const trackerLink = tracker.use(
|
||||||
|
createRelayMiddleware((variables) => {
|
||||||
|
return {
|
||||||
|
...variables,
|
||||||
|
password: '***',
|
||||||
|
};
|
||||||
|
}),
|
||||||
|
);
|
||||||
```
|
```
|
||||||
|
|
||||||
Or you can manually put `recordGraphQL` call
|
Or you can manually put `recordGraphQL` call
|
||||||
|
|
@ -52,22 +65,22 @@ then you should do something like below
|
||||||
import { createGraphqlMiddleware } from '@openreplay/tracker-graphql'; // see above for recordGraphQL definition
|
import { createGraphqlMiddleware } from '@openreplay/tracker-graphql'; // see above for recordGraphQL definition
|
||||||
import { Environment } from 'relay-runtime';
|
import { Environment } from 'relay-runtime';
|
||||||
|
|
||||||
const handler = createGraphqlMiddleware(tracker)
|
const handler = tracker.use(createGraphqlMiddleware());
|
||||||
|
|
||||||
function fetchQuery(operation, variables, cacheConfig, uploadables) {
|
function fetchQuery(operation, variables, cacheConfig, uploadables) {
|
||||||
return fetch('www.myapi.com/resource', {
|
return fetch('www.myapi.com/resource', {
|
||||||
// ...
|
// ...
|
||||||
})
|
})
|
||||||
.then(response => response.json())
|
.then((response) => response.json())
|
||||||
.then(result =>
|
.then((result) =>
|
||||||
handler(
|
handler(
|
||||||
// op kind, name, variables, response, duration (default 0)
|
// op kind, name, variables, response, duration (default 0)
|
||||||
operation.operationKind,
|
operation.operationKind,
|
||||||
operation.name,
|
operation.name,
|
||||||
variables,
|
variables,
|
||||||
result,
|
result,
|
||||||
duration,
|
duration,
|
||||||
),
|
),
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -81,10 +94,23 @@ See [Relay Network Layer](https://relay.dev/docs/en/network-layer) for details.
|
||||||
For [Apollo](https://www.apollographql.com/) you should create a new `ApolloLink`
|
For [Apollo](https://www.apollographql.com/) you should create a new `ApolloLink`
|
||||||
|
|
||||||
```js
|
```js
|
||||||
import { createTrackerLink } from '@openreplay/tracker-graphql'
|
import { createTrackerLink } from '@openreplay/tracker-graphql';
|
||||||
|
|
||||||
const trackerLink = createTrackerLink(tracker);
|
const trackerLink = tracker.use(createTrackerLink());
|
||||||
const yourLink = new ApolloLink(trackerLink)
|
const yourLink = new ApolloLink(trackerLink);
|
||||||
|
```
|
||||||
|
|
||||||
|
You can pass a Sanitizer function to `createRelayMiddleware` to sanitize the variables and data before sending them to OpenReplay.
|
||||||
|
|
||||||
|
```js
|
||||||
|
const trackerLink = tracker.use(
|
||||||
|
createTrackerLink((variables) => {
|
||||||
|
return {
|
||||||
|
...variables,
|
||||||
|
password: '***',
|
||||||
|
};
|
||||||
|
}),
|
||||||
|
);
|
||||||
```
|
```
|
||||||
|
|
||||||
Alternatively you can use generic graphql handler:
|
Alternatively you can use generic graphql handler:
|
||||||
|
|
@ -93,18 +119,21 @@ Alternatively you can use generic graphql handler:
|
||||||
import { createGraphqlMiddleware } from '@openreplay/tracker-graphql'; // see above for recordGraphQL definition
|
import { createGraphqlMiddleware } from '@openreplay/tracker-graphql'; // see above for recordGraphQL definition
|
||||||
import { ApolloLink } from 'apollo-link';
|
import { ApolloLink } from 'apollo-link';
|
||||||
|
|
||||||
const handler = createGraphqlMiddleware(tracker)
|
const handler = tracker.use(createGraphqlMiddleware());
|
||||||
|
|
||||||
const trackerApolloLink = new ApolloLink((operation, forward) => {
|
const trackerApolloLink = new ApolloLink((operation, forward) => {
|
||||||
return forward(operation).map(result =>
|
operation.setContext({ start: performance.now() });
|
||||||
handler(
|
return forward(operation).map((result) => {
|
||||||
|
const time = performance.now() - operation.getContext().start;
|
||||||
|
return handler(
|
||||||
// op kind, name, variables, response, duration (default 0)
|
// op kind, name, variables, response, duration (default 0)
|
||||||
operation.query.definitions[0].operation,
|
operation.query.definitions[0].operation,
|
||||||
operation.operationName,
|
operation.operationName,
|
||||||
operation.variables,
|
operation.variables,
|
||||||
result,
|
result,
|
||||||
),
|
time,
|
||||||
);
|
);
|
||||||
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
const link = ApolloLink.from([
|
const link = ApolloLink.from([
|
||||||
|
|
|
||||||
|
|
@ -1,5 +1,6 @@
|
||||||
import { App, Messages } from '@openreplay/tracker';
|
import { App, Messages } from '@openreplay/tracker';
|
||||||
import Observable from 'zen-observable';
|
import Observable from 'zen-observable';
|
||||||
|
import { Sanitizer } from './types';
|
||||||
|
|
||||||
type Operation = {
|
type Operation = {
|
||||||
query: Record<string, any>;
|
query: Record<string, any>;
|
||||||
|
|
@ -9,48 +10,63 @@ type Operation = {
|
||||||
};
|
};
|
||||||
type NextLink = (operation: Operation) => Observable<Record<string, any>>;
|
type NextLink = (operation: Operation) => Observable<Record<string, any>>;
|
||||||
|
|
||||||
export const createTrackerLink = (app: App | null) => {
|
export const createTrackerLink = (
|
||||||
if (!app) {
|
sanitizer?: Sanitizer<Record<string, any> | undefined | null>,
|
||||||
return (operation: Operation, forward: NextLink) => forward(operation);
|
) => {
|
||||||
}
|
return (app: App | null) => {
|
||||||
return (operation: Operation, forward: NextLink) => {
|
if (!app) {
|
||||||
return new Observable((observer) => {
|
return (operation: Operation, forward: NextLink) => forward(operation);
|
||||||
const start = app.timestamp();
|
}
|
||||||
const observable = forward(operation);
|
return (operation: Operation, forward: NextLink) => {
|
||||||
const subscription = observable.subscribe({
|
return new Observable((observer) => {
|
||||||
next(value) {
|
const start = app.timestamp();
|
||||||
const end = app.timestamp();
|
const observable = forward(operation);
|
||||||
app.send(
|
const subscription = observable.subscribe({
|
||||||
Messages.GraphQL(
|
next(value) {
|
||||||
operation.query.definitions[0].kind,
|
const end = app.timestamp();
|
||||||
operation.operationName,
|
const operationDefinition = operation.query.definitions[0];
|
||||||
JSON.stringify(operation.variables),
|
app.send(
|
||||||
JSON.stringify(value.data),
|
Messages.GraphQL(
|
||||||
end - start,
|
operationDefinition.kind === 'OperationDefinition'
|
||||||
),
|
? operationDefinition.operation
|
||||||
);
|
: 'unknown?',
|
||||||
observer.next(value);
|
operation.operationName,
|
||||||
},
|
JSON.stringify(
|
||||||
error(error) {
|
sanitizer
|
||||||
const end = app.timestamp();
|
? sanitizer(operation.variables)
|
||||||
app.send(
|
: operation.variables,
|
||||||
Messages.GraphQL(
|
),
|
||||||
operation.query.definitions[0].kind,
|
JSON.stringify(sanitizer ? sanitizer(value.data) : value.data),
|
||||||
operation.operationName,
|
end - start,
|
||||||
JSON.stringify(operation.variables),
|
),
|
||||||
JSON.stringify(error),
|
);
|
||||||
end - start,
|
observer.next(value);
|
||||||
),
|
},
|
||||||
);
|
error(error) {
|
||||||
observer.error(error);
|
const end = app.timestamp();
|
||||||
},
|
app.send(
|
||||||
complete() {
|
Messages.GraphQL(
|
||||||
observer.complete();
|
operation.query.definitions[0].kind,
|
||||||
},
|
operation.operationName,
|
||||||
});
|
JSON.stringify(
|
||||||
|
sanitizer
|
||||||
|
? sanitizer(operation.variables)
|
||||||
|
: operation.variables,
|
||||||
|
),
|
||||||
|
JSON.stringify(error),
|
||||||
|
end - start,
|
||||||
|
),
|
||||||
|
);
|
||||||
|
observer.error(error);
|
||||||
|
},
|
||||||
|
complete() {
|
||||||
|
observer.complete();
|
||||||
|
},
|
||||||
|
});
|
||||||
|
|
||||||
return () => subscription.unsubscribe();
|
return () => subscription.unsubscribe();
|
||||||
});
|
});
|
||||||
|
};
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -1,4 +1,4 @@
|
||||||
import { App, Messages } from "@openreplay/tracker";
|
import { App, Messages } from '@openreplay/tracker';
|
||||||
|
|
||||||
function createGraphqlMiddleware() {
|
function createGraphqlMiddleware() {
|
||||||
return (app: App | null) => {
|
return (app: App | null) => {
|
||||||
|
|
@ -10,7 +10,7 @@ function createGraphqlMiddleware() {
|
||||||
operationName: string,
|
operationName: string,
|
||||||
variables: any,
|
variables: any,
|
||||||
result: any,
|
result: any,
|
||||||
duration = 0
|
duration = 0,
|
||||||
) => {
|
) => {
|
||||||
try {
|
try {
|
||||||
app.send(
|
app.send(
|
||||||
|
|
@ -30,4 +30,4 @@ function createGraphqlMiddleware() {
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
export default createGraphqlMiddleware
|
export default createGraphqlMiddleware;
|
||||||
|
|
|
||||||
|
|
@ -1,9 +1,11 @@
|
||||||
import createTrackerLink from './apolloMiddleware.js';
|
import createTrackerLink from './apolloMiddleware.js';
|
||||||
import createRelayMiddleware from './relayMiddleware.js';
|
import createRelayMiddleware from './relayMiddleware.js';
|
||||||
import createGraphqlMiddleware from './graphqlMiddleware.js';
|
import createGraphqlMiddleware from './graphqlMiddleware.js';
|
||||||
|
import { Sanitizer } from './types.js';
|
||||||
|
|
||||||
export {
|
export {
|
||||||
createTrackerLink,
|
createTrackerLink,
|
||||||
createRelayMiddleware,
|
createRelayMiddleware,
|
||||||
createGraphqlMiddleware,
|
createGraphqlMiddleware,
|
||||||
}
|
Sanitizer,
|
||||||
|
};
|
||||||
|
|
|
||||||
|
|
@ -1,37 +1,55 @@
|
||||||
import { App, Messages } from '@openreplay/tracker';
|
import { App, Messages } from '@openreplay/tracker';
|
||||||
import type { Middleware, RelayRequest } from './relaytypes';
|
import type { Middleware, RelayRequest } from './relaytypes';
|
||||||
|
import { Sanitizer } from './types';
|
||||||
|
|
||||||
const createRelayMiddleware = (app: App | null): Middleware => {
|
const createRelayMiddleware = (sanitizer?: Sanitizer<Record<string, any>>) => {
|
||||||
if (!app) {
|
return (app: App | null): Middleware => {
|
||||||
return (next) => async (req) => await next(req);
|
if (!app) {
|
||||||
}
|
return (next) => async (req) => await next(req);
|
||||||
return (next) => async (req) => {
|
|
||||||
const start = app.timestamp();
|
|
||||||
const resp = await next(req)
|
|
||||||
const end = app.timestamp();
|
|
||||||
if ('requests' in req) {
|
|
||||||
req.requests.forEach((request) => {
|
|
||||||
app.send(getMessage(request, resp.json as Record<string, any>, end - start))
|
|
||||||
})
|
|
||||||
} else {
|
|
||||||
app.send(getMessage(req, resp.json as Record<string, any>, end - start))
|
|
||||||
}
|
}
|
||||||
return resp;
|
return (next) => async (req) => {
|
||||||
}
|
const start = app.timestamp();
|
||||||
|
const resp = await next(req);
|
||||||
|
const end = app.timestamp();
|
||||||
|
if ('requests' in req) {
|
||||||
|
req.requests.forEach((request) => {
|
||||||
|
app.send(
|
||||||
|
getMessage(
|
||||||
|
request,
|
||||||
|
resp.json as Record<string, any>,
|
||||||
|
end - start,
|
||||||
|
sanitizer,
|
||||||
|
),
|
||||||
|
);
|
||||||
|
});
|
||||||
|
} else {
|
||||||
|
app.send(
|
||||||
|
getMessage(
|
||||||
|
req,
|
||||||
|
resp.json as Record<string, any>,
|
||||||
|
end - start,
|
||||||
|
sanitizer,
|
||||||
|
),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
return resp;
|
||||||
|
};
|
||||||
|
};
|
||||||
};
|
};
|
||||||
|
|
||||||
function getMessage(request: RelayRequest, json: Record<string, any>, duration: number) {
|
function getMessage(
|
||||||
|
request: RelayRequest,
|
||||||
|
json: Record<string, any>,
|
||||||
|
duration: number,
|
||||||
|
sanitizer?: Sanitizer<Record<string, any>>,
|
||||||
|
) {
|
||||||
const opKind = request.operation.kind;
|
const opKind = request.operation.kind;
|
||||||
const opName = request.operation.name;
|
const opName = request.operation.name;
|
||||||
const vars = JSON.stringify(request.variables)
|
const vars = JSON.stringify(
|
||||||
const opResp = JSON.stringify(json)
|
sanitizer ? sanitizer(request.variables) : request.variables,
|
||||||
return Messages.GraphQL(
|
);
|
||||||
opKind,
|
const opResp = JSON.stringify(sanitizer ? sanitizer(json) : json);
|
||||||
opName,
|
return Messages.GraphQL(opKind, opName, vars, opResp, duration);
|
||||||
vars,
|
|
||||||
opResp,
|
|
||||||
duration
|
|
||||||
)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
export default createRelayMiddleware
|
export default createRelayMiddleware;
|
||||||
|
|
|
||||||
|
|
@ -1,4 +1,3 @@
|
||||||
|
|
||||||
type ConcreteBatch = {
|
type ConcreteBatch = {
|
||||||
kind: 'Batch';
|
kind: 'Batch';
|
||||||
fragment: any;
|
fragment: any;
|
||||||
|
|
@ -9,7 +8,7 @@ type ConcreteBatch = {
|
||||||
text: string | null;
|
text: string | null;
|
||||||
operationKind: string;
|
operationKind: string;
|
||||||
};
|
};
|
||||||
type Variables = { [name: string]: any };
|
export type Variables = { [name: string]: any };
|
||||||
interface FetchOpts {
|
interface FetchOpts {
|
||||||
url?: string;
|
url?: string;
|
||||||
method: 'POST' | 'GET';
|
method: 'POST' | 'GET';
|
||||||
|
|
@ -17,7 +16,13 @@ interface FetchOpts {
|
||||||
body: string | FormData;
|
body: string | FormData;
|
||||||
credentials?: 'same-origin' | 'include' | 'omit';
|
credentials?: 'same-origin' | 'include' | 'omit';
|
||||||
mode?: 'cors' | 'websocket' | 'navigate' | 'no-cors' | 'same-origin';
|
mode?: 'cors' | 'websocket' | 'navigate' | 'no-cors' | 'same-origin';
|
||||||
cache?: 'default' | 'no-store' | 'reload' | 'no-cache' | 'force-cache' | 'only-if-cached';
|
cache?:
|
||||||
|
| 'default'
|
||||||
|
| 'no-store'
|
||||||
|
| 'reload'
|
||||||
|
| 'no-cache'
|
||||||
|
| 'force-cache'
|
||||||
|
| 'only-if-cached';
|
||||||
redirect?: 'follow' | 'error' | 'manual';
|
redirect?: 'follow' | 'error' | 'manual';
|
||||||
signal?: AbortSignal;
|
signal?: AbortSignal;
|
||||||
[name: string]: any;
|
[name: string]: any;
|
||||||
|
|
|
||||||
1
tracker/tracker-graphql/src/types.ts
Normal file
1
tracker/tracker-graphql/src/types.ts
Normal file
|
|
@ -0,0 +1 @@
|
||||||
|
export type Sanitizer<T> = (values: T) => Partial<T>;
|
||||||
|
|
@ -963,8 +963,8 @@ export default class App {
|
||||||
deviceMemory,
|
deviceMemory,
|
||||||
jsHeapSizeLimit,
|
jsHeapSizeLimit,
|
||||||
timezone: getTimezone(),
|
timezone: getTimezone(),
|
||||||
width: window.innerWidth,
|
width: window.screen.width,
|
||||||
height: window.innerHeight,
|
height: window.screen.height,
|
||||||
}),
|
}),
|
||||||
})
|
})
|
||||||
const {
|
const {
|
||||||
|
|
@ -1220,7 +1220,9 @@ export default class App {
|
||||||
timezone: getTimezone(),
|
timezone: getTimezone(),
|
||||||
condition: conditionName,
|
condition: conditionName,
|
||||||
assistOnly: startOpts.assistOnly ?? this.socketMode,
|
assistOnly: startOpts.assistOnly ?? this.socketMode,
|
||||||
}),
|
width: window.screen.width,
|
||||||
|
height: window.screen.height
|
||||||
|
}),
|
||||||
})
|
})
|
||||||
if (r.status !== 200) {
|
if (r.status !== 200) {
|
||||||
const error = await r.text()
|
const error = await r.text()
|
||||||
|
|
|
||||||
Loading…
Add table
Reference in a new issue