Compare commits

...
Sign in to create a new pull request.

21 commits

Author SHA1 Message Date
Mehdi Osman
dda7affd5d
Added mobile and canvas related topics 2024-07-24 14:19:55 -04:00
Atef Ben Ali
6e2a772e7f
docs: update README_AR.md file (#2421) 2024-07-24 09:44:21 -04:00
Mehdi Osman
dcce3569fb
Increment chalice chart version (#2417)
Co-authored-by: GitHub Action <action@github.com>
2024-07-23 13:17:48 +02:00
Kraiem Taha Yassine
867247dbc0
fix(chalice): fixed insights with filter steps (#2416) 2024-07-23 13:01:48 +02:00
PiR
754293e29d
Tracker GrahpQL: update doc and tracker initialization + add option to pass sanitizer function (#2402)
* fix(graphQL): update doc and tracker initialization + add option to pass sanitizer function

* improvement(graphQL): improve sanitizer type & apollo operation name
2024-07-22 16:07:12 +02:00
Mehdi Osman
ddd037ce79
Increment chalice chart version (#2411)
Co-authored-by: GitHub Action <action@github.com>
2024-07-19 15:09:47 +02:00
Kraiem Taha Yassine
66f4c5c93b
fix(chalice): stop SA from logout (#2410) 2024-07-19 15:03:59 +02:00
Mehdi Osman
66e4d133ad
Increment chalice chart version (#2407)
Co-authored-by: GitHub Action <action@github.com>
2024-07-18 13:17:27 +02:00
Kraiem Taha Yassine
f9f8853ab0
fix(chalice): fixed search mobile sessions (#2406)
fix(chalice): fixed autocomplete mobile sessions
2024-07-18 13:08:05 +02:00
Mehdi Osman
e0bb6fea9d
Updated patch build from main 4e7efaecde (#2405)
* Increment frontend chart version

* Increment db chart version

---------

Co-authored-by: GitHub Action <action@github.com>
2024-07-18 12:46:42 +02:00
Alexander
4e7efaecde
Heatmaps fix (float coordinates) (#2403) (#2404)
* feat(spot): use float click coordinates instead of ints in PG

* feat(db): added support for float clicks in CH

* feat(db): fix float instead of uint8

* feat(mobile): new naming for mobile autocomplete types
2024-07-18 12:38:47 +02:00
Delirium
54a9624332
Heatmaps patch 2 (#2400)
* fix ui: move clickmap overlay inside replay vdom, refactor renderer scaling

* fix ui: fix first event calculation
2024-07-17 18:57:21 +02:00
Mehdi Osman
1ddffca572
Increment frontend chart version (#2395)
Co-authored-by: GitHub Action <action@github.com>
2024-07-16 17:38:25 +02:00
Delirium
c91881413a
fix manager event reads for mobile (#2394) 2024-07-16 17:34:12 +02:00
Mehdi Osman
ba2d9eb81c
Increment chalice chart version (#2393)
Co-authored-by: GitHub Action <action@github.com>
2024-07-16 17:20:59 +02:00
Kraiem Taha Yassine
c845415e1e
Patch/api v1.19.0 (#2392)
* fix(chalice): reversed count&total for card-tables to confuse devs

* fix(DB): changed normalized_x&y col-type
2024-07-16 17:15:38 +02:00
Mehdi Osman
ee0ede8478
Increment chalice chart version (#2391)
Co-authored-by: GitHub Action <action@github.com>
2024-07-16 14:26:47 +02:00
Kraiem Taha Yassine
72afae226b
fix(chalice): fixed missing totalSessions in card-tables in EE (#2390)
* fix(chalice): fixed missing totalSessions in card-tables in EE

* fix(chalice): fixed missing totalSessions in card-tables in EE
2024-07-16 14:16:26 +02:00
Shekar Siri
b3f545849a
fix(ui): use count instead of totalSessions (#2387) 2024-07-12 17:38:44 +02:00
Mehdi Osman
cd2966fb9f
Increment chalice chart version (#2384)
Co-authored-by: GitHub Action <action@github.com>
2024-07-11 11:50:42 +02:00
Kraiem Taha Yassine
4b91dcded0
Patch/api v1.19.0 (#2383)
* fix(chalice): fixed create heatmap card EE

* fix(chalice): fixed click_rage-heatmap card EE

* fix(chalice): fixed click_rage-heatmap ambiguous alias EE
2024-07-11 11:36:07 +02:00
41 changed files with 457 additions and 245 deletions

View file

@ -55,17 +55,17 @@ OpenReplay هو مجموعة إعادة تشغيل الجلسة التي يمك
## الميزات ## الميزات
- **إعادة تشغيل الجلسة:** تتيح لك إعادة تشغيل الجلسة إعادة عيش تجربة مستخدميك، ورؤية أين يواجهون صعوبة وكيف يؤثر ذلك على سلوكهم. يتم تحليل كل إعادة تشغيل للجلسة تلقائيًا بناءً على الأساليب الاستدلالية، لسهولة التقييم. - **إعادة تشغيل الجلسة:** تتيح لك إعادة تشغيل الجلسة إعادة عيش تجربة مستخدميك، ورؤية أين يواجهون صعوبة وكيف يؤثر ذلك على سلوكهم. يتم تحليل كل إعادة تشغيل للجلسة تلقائيًا بناءً على الأساليب الاستدلالية، لسهولة التقييم.
- **أدوات التطوير (DevTools):** إنها مثل التصحيح في متصفحك الخاص. يوفر لك OpenReplay السياق الكامل (نشاط الشبكة، أخطاء JavaScript، إجراءات/حالة التخزين وأكثر من 40 مقياسًا) حتى تتمكن من إعادة إنتاج الأخطاء فورًا وفهم مشكلات الأداء. - **أدوات التطوير (DevTools):** إنها مثل المصحح (debugger) في متصفحك الخاص. يوفر لك OpenReplay السياق الكامل (نشاط الشبكة، أخطاء JavaScript، إجراءات/حالة التخزين وأكثر من 40 مقياسًا) حتى تتمكن من إعادة إنتاج الأخطاء فورًا وفهم مشكلات الأداء.
- **المساعدة (Assist):** تساعدك في دعم مستخدميك من خلال رؤية شاشتهم مباشرة والانضمام فورًا إلى مكالمة (WebRTC) معهم دون الحاجة إلى برامج مشاركة الشاشة من جهات خارجية. - **المساعدة (Assist):** تساعدك في دعم مستخدميك من خلال رؤية شاشتهم مباشرة والانضمام فورًا إلى مكالمة (WebRTC) معهم دون الحاجة إلى برامج مشاركة الشاشة من جهات خارجية.
- **البحث الشامل (Omni-search):** ابحث وفرز حسب أي عملية/معيار للمستخدم تقريبًا، أو سمة الجلسة أو الحدث التقني، حتى تتمكن من الرد على أي سؤال. لا يلزم تجهيز. - **البحث الشامل (Omni-search):** ابحث وافرز حسب أي عملية/معيار للمستخدم تقريبًا، أو سمة الجلسة أو الحدث التقني، حتى تتمكن من الرد على أي سؤال. لا يلزم تجهيز.
- **الأنفاق (Funnels):** للكشف عن المشكلات الأكثر تأثيرًا التي تسبب في فقدان التحويل والإيرادات. - **الأنفاق (Funnels):** للكشف عن المشكلات الأكثر تأثيرًا التي تسبب في فقدان التحويل والإيرادات.
- **ضوابط الخصوصية الدقيقة:** اختر ماذا تريد التقاطه، ماذا تريد أن تخفي أو تجاهل حتى لا تصل بيانات المستخدم حتى إلى خوادمك. - **ضوابط الخصوصية الدقيقة:** اختر ماذا تريد التقاطه، ماذا تريد أن تخفي أو تتجاهل حتى لا تصل بيانات المستخدم حتى إلى خوادمك.
- **موجهة للمكونات الإضافية (Plugins oriented):** تصل إلى السبب الجذري بشكل أسرع عن طريق تتبع حالة التطبيق (Redux، VueX، MobX، NgRx، Pinia، وZustand) وتسجيل استعلامات GraphQL (Apollo، Relay) وطلبات Fetch/Axios. - **موجهة للمكونات الإضافية (Plugins oriented):** يمكنك الوصول إلى السبب الجذري بشكل أسرع عن طريق تتبع حالة التطبيق (Redux، VueX، MobX، NgRx، Pinia، وZustand) وتسجيل استعلامات GraphQL (Apollo، Relay) وطلبات Fetch/Axios.
- **التكاملات (Integrations):** مزامنة سجلات الخادم الخلفي مع إعادات التشغيل للجلسات ورؤية ما حدث من الأمام إلى الخلف. يدعم OpenReplay Sentry وDatadog وCloudWatch وStackdriver وElastic والمزيد. - **التكاملات (Integrations):** مزامنة سجلات الخادم الخلفي مع إعادات التشغيل للجلسات ورؤية ما حدث من الأمام إلى الخلف. يدعم OpenReplay Sentry وDatadog وCloudWatch وStackdriver وElastic والمزيد.
## خيارات النشر ## خيارات النشر
يمكن نشر OpenReplay في أي مكان. اتبع دليلنا الخطوة بالخطوة لنشره على خدمات السحابة العامة الرئيسية: يمكن نشر OpenReplay في أي مكان. اتبع دليلنا خطوة بخطوة لنشره على خدمات السحابة العامة الرئيسة:
- [AWS](https://docs.openreplay.com/deployment/deploy-aws) - [AWS](https://docs.openreplay.com/deployment/deploy-aws)
- [Google Cloud](https://docs.openreplay.com/deployment/deploy-gcp) - [Google Cloud](https://docs.openreplay.com/deployment/deploy-gcp)

View file

@ -319,13 +319,14 @@ def create_card(project_id, user_id, data: schemas.CardSchema, dashboard=False):
session_data = None session_data = None
if data.metric_type == schemas.MetricType.heat_map: if data.metric_type == schemas.MetricType.heat_map:
if data.session_id is not None: if data.session_id is not None:
session_data = json.dumps({"sessionId": data.session_id}) session_data = {"sessionId": data.session_id}
else: else:
session_data = __get_heat_map_chart(project_id=project_id, user_id=user_id, session_data = __get_heat_map_chart(project_id=project_id, user_id=user_id,
data=data, include_mobs=False) data=data, include_mobs=False)
if session_data is not None: if session_data is not None:
session_data = json.dumps({"sessionId": session_data["sessionId"]}) session_data = {"sessionId": session_data["sessionId"]}
_data = {"session_data": session_data}
_data = {"session_data": json.dumps(session_data) if session_data is not None else None}
for i, s in enumerate(data.series): for i, s in enumerate(data.series):
for k in s.model_dump().keys(): for k in s.model_dump().keys():
_data[f"{k}_{i}"] = s.__getattribute__(k) _data[f"{k}_{i}"] = s.__getattribute__(k)

View file

@ -359,12 +359,12 @@ def search2_table(data: schemas.SessionsSearchPayloadSchema, project_id: int, de
distinct_on += ",path" distinct_on += ",path"
if metric_format == schemas.MetricExtendedFormatType.session_count: if metric_format == schemas.MetricExtendedFormatType.session_count:
main_query = f"""SELECT COUNT(*) AS count, main_query = f"""SELECT COUNT(*) AS count,
COALESCE(SUM(users_sessions.session_count),0) AS total_sessions, COALESCE(SUM(users_sessions.session_count),0) AS count,
COALESCE(JSONB_AGG(users_sessions) COALESCE(JSONB_AGG(users_sessions)
FILTER ( WHERE rn > %(limit_s)s FILTER ( WHERE rn > %(limit_s)s
AND rn <= %(limit_e)s ), '[]'::JSONB) AS values AND rn <= %(limit_e)s ), '[]'::JSONB) AS values
FROM (SELECT {main_col} AS name, FROM (SELECT {main_col} AS name,
count(DISTINCT session_id) AS session_count, count(DISTINCT session_id) AS total,
ROW_NUMBER() OVER (ORDER BY count(full_sessions) DESC) AS rn ROW_NUMBER() OVER (ORDER BY count(full_sessions) DESC) AS rn
FROM (SELECT * FROM (SELECT *
FROM (SELECT DISTINCT ON({distinct_on}) s.session_id, s.user_uuid, FROM (SELECT DISTINCT ON({distinct_on}) s.session_id, s.user_uuid,
@ -379,7 +379,7 @@ def search2_table(data: schemas.SessionsSearchPayloadSchema, project_id: int, de
ORDER BY session_count DESC) AS users_sessions;""" ORDER BY session_count DESC) AS users_sessions;"""
else: else:
main_query = f"""SELECT COUNT(*) AS count, main_query = f"""SELECT COUNT(*) AS count,
COALESCE(SUM(users_sessions.user_count),0) AS total_users, COALESCE(SUM(users_sessions.user_count),0) AS count,
COALESCE(JSONB_AGG(users_sessions) FILTER ( WHERE rn <= 200 ), '[]'::JSONB) AS values COALESCE(JSONB_AGG(users_sessions) FILTER ( WHERE rn <= 200 ), '[]'::JSONB) AS values
FROM (SELECT {main_col} AS name, FROM (SELECT {main_col} AS name,
count(DISTINCT user_id) AS user_count, count(DISTINCT user_id) AS user_count,
@ -420,12 +420,12 @@ def search_table_of_individual_issues(data: schemas.SessionsSearchPayloadSchema,
full_args["issues_limit_s"] = (data.page - 1) * data.limit full_args["issues_limit_s"] = (data.page - 1) * data.limit
full_args["issues_limit_e"] = data.page * data.limit full_args["issues_limit_e"] = data.page * data.limit
main_query = cur.mogrify(f"""SELECT COUNT(1) AS count, main_query = cur.mogrify(f"""SELECT COUNT(1) AS count,
COALESCE(SUM(session_count), 0) AS total_sessions, COALESCE(SUM(session_count), 0) AS count,
COALESCE(JSONB_AGG(ranked_issues) COALESCE(JSONB_AGG(ranked_issues)
FILTER ( WHERE rn > %(issues_limit_s)s FILTER ( WHERE rn > %(issues_limit_s)s
AND rn <= %(issues_limit_e)s ), '[]'::JSONB) AS values AND rn <= %(issues_limit_e)s ), '[]'::JSONB) AS values
FROM (SELECT *, ROW_NUMBER() OVER (ORDER BY session_count DESC) AS rn FROM (SELECT *, ROW_NUMBER() OVER (ORDER BY session_count DESC) AS rn
FROM (SELECT type AS name, context_string AS value, COUNT(DISTINCT session_id) AS session_count FROM (SELECT type AS name, context_string AS value, COUNT(DISTINCT session_id) AS total
FROM (SELECT session_id FROM (SELECT session_id
{query_part}) AS filtered_sessions {query_part}) AS filtered_sessions
INNER JOIN events_common.issues USING (session_id) INNER JOIN events_common.issues USING (session_id)
@ -814,12 +814,6 @@ def search_query_parts(data: schemas.SessionsSearchPayloadSchema, error_status,
event_where.append( event_where.append(
sh.multi_conditions(f"main.{events.EventType.VIEW_MOBILE.column} {op} %({e_k})s", sh.multi_conditions(f"main.{events.EventType.VIEW_MOBILE.column} {op} %({e_k})s",
event.value, value_key=e_k)) event.value, value_key=e_k))
elif event_type == events.EventType.SWIPE_MOBILE.ui_type and platform == "ios":
event_from = event_from % f"{events.EventType.SWIPE_MOBILE.table} AS main "
if not is_any:
event_where.append(
sh.multi_conditions(f"main.{events.EventType.SWIPE_MOBILE.column} {op} %({e_k})s",
event.value, value_key=e_k))
elif event_type == events.EventType.CUSTOM.ui_type: elif event_type == events.EventType.CUSTOM.ui_type:
event_from = event_from % f"{events.EventType.CUSTOM.table} AS main " event_from = event_from % f"{events.EventType.CUSTOM.table} AS main "
if not is_any: if not is_any:
@ -855,7 +849,7 @@ def search_query_parts(data: schemas.SessionsSearchPayloadSchema, error_status,
event_where.append(sh.multi_conditions(f"main1.source = %({s_k})s", event.source, value_key=s_k)) event_where.append(sh.multi_conditions(f"main1.source = %({s_k})s", event.source, value_key=s_k))
# ----- IOS # ----- Mobile
elif event_type == events.EventType.CLICK_MOBILE.ui_type: elif event_type == events.EventType.CLICK_MOBILE.ui_type:
event_from = event_from % f"{events.EventType.CLICK_MOBILE.table} AS main " event_from = event_from % f"{events.EventType.CLICK_MOBILE.table} AS main "
if not is_any: if not is_any:
@ -897,6 +891,13 @@ def search_query_parts(data: schemas.SessionsSearchPayloadSchema, error_status,
event_where.append( event_where.append(
sh.multi_conditions(f"(main1.reason {op} %({e_k})s OR main1.name {op} %({e_k})s)", sh.multi_conditions(f"(main1.reason {op} %({e_k})s OR main1.name {op} %({e_k})s)",
event.value, value_key=e_k)) event.value, value_key=e_k))
elif event_type == events.EventType.SWIPE_MOBILE.ui_type and platform != "web":
event_from = event_from % f"{events.EventType.SWIPE_MOBILE.table} AS main "
if not is_any:
event_where.append(
sh.multi_conditions(f"main.{events.EventType.SWIPE_MOBILE.column} {op} %({e_k})s",
event.value, value_key=e_k))
elif event_type == schemas.PerformanceEventType.fetch_failed: elif event_type == schemas.PerformanceEventType.fetch_failed:
event_from = event_from % f"{events.EventType.REQUEST.table} AS main " event_from = event_from % f"{events.EventType.REQUEST.table} AS main "
if not is_any: if not is_any:

View file

@ -19,6 +19,7 @@ from routers.base import get_routers
public_app, app, app_apikey = get_routers() public_app, app, app_apikey = get_routers()
@app.get('/{projectId}/autocomplete', tags=["events"])
@app.get('/{projectId}/events/search', tags=["events"]) @app.get('/{projectId}/events/search', tags=["events"])
def events_search(projectId: int, q: str, def events_search(projectId: int, q: str,
type: Union[schemas.FilterType, schemas.EventType, type: Union[schemas.FilterType, schemas.EventType,

View file

@ -471,13 +471,13 @@ class EventType(str, Enum):
state_action = "stateAction" state_action = "stateAction"
error = "error" error = "error"
tag = "tag" tag = "tag"
click_mobile = "tapIos" click_mobile = "click_mobile"
input_mobile = "inputIos" input_mobile = "input_mobile"
view_mobile = "viewIos" view_mobile = "view_mobile"
custom_mobile = "customIos" custom_mobile = "custom_mobile"
request_mobile = "requestIos" request_mobile = "request_mobile"
error_mobile = "errorIos" error_mobile = "error_mobile"
swipe_mobile = "swipeIos" swipe_mobile = "swipe_mobile"
class PerformanceEventType(str, Enum): class PerformanceEventType(str, Enum):

View file

@ -81,13 +81,13 @@ func (s *saverImpl) handleMobileMessage(msg Message) error {
if err = s.sessions.UpdateUserID(session.SessionID, m.ID); err != nil { if err = s.sessions.UpdateUserID(session.SessionID, m.ID); err != nil {
return err return err
} }
s.pg.InsertAutocompleteValue(session.SessionID, session.ProjectID, "USERID_Mobile", m.ID) s.pg.InsertAutocompleteValue(session.SessionID, session.ProjectID, "USERID_MOBILE", m.ID)
return nil return nil
case *MobileUserAnonymousID: case *MobileUserAnonymousID:
if err = s.sessions.UpdateAnonymousID(session.SessionID, m.ID); err != nil { if err = s.sessions.UpdateAnonymousID(session.SessionID, m.ID); err != nil {
return err return err
} }
s.pg.InsertAutocompleteValue(session.SessionID, session.ProjectID, "USERANONYMOUSID_Mobile", m.ID) s.pg.InsertAutocompleteValue(session.SessionID, session.ProjectID, "USERANONYMOUSID_MOBILE", m.ID)
return nil return nil
case *MobileMetadata: case *MobileMetadata:
return s.sessions.UpdateMetadata(m.SessionID(), m.Key, m.Value) return s.sessions.UpdateMetadata(m.SessionID(), m.Key, m.Value)

View file

@ -132,8 +132,15 @@ func (conn *Conn) InsertWebClickEvent(sess *sessions.Session, e *messages.MouseC
} }
var host, path string var host, path string
host, path, _, _ = url.GetURLParts(e.Url) host, path, _, _ = url.GetURLParts(e.Url)
if e.NormalizedX <= 100 && e.NormalizedY <= 100 { if e.NormalizedX != 101 && e.NormalizedY != 101 {
if err := conn.bulks.Get("webClickXYEvents").Append(sess.SessionID, truncSqIdx(e.MsgID()), e.Timestamp, e.Label, e.Selector, host+path, path, e.HesitationTime, e.NormalizedX, e.NormalizedY); err != nil { // To support previous versions of tracker
if e.NormalizedX <= 100 && e.NormalizedY <= 100 {
e.NormalizedX *= 100
e.NormalizedY *= 100
}
normalizedX := float32(e.NormalizedX) / 100.0
normalizedY := float32(e.NormalizedY) / 100.0
if err := conn.bulks.Get("webClickXYEvents").Append(sess.SessionID, truncSqIdx(e.MsgID()), e.Timestamp, e.Label, e.Selector, host+path, path, e.HesitationTime, normalizedX, normalizedY); err != nil {
sessCtx := context.WithValue(context.Background(), "sessionID", sess.SessionID) sessCtx := context.WithValue(context.Background(), "sessionID", sess.SessionID)
conn.log.Error(sessCtx, "insert web click event in bulk err: %s", err) conn.log.Error(sessCtx, "insert web click event in bulk err: %s", err)
} }

View file

@ -13,14 +13,14 @@ func (conn *Conn) InsertMobileEvent(session *sessions.Session, e *messages.Mobil
if err := conn.InsertCustomEvent(session.SessionID, e.Timestamp, truncSqIdx(e.Index), e.Name, e.Payload); err != nil { if err := conn.InsertCustomEvent(session.SessionID, e.Timestamp, truncSqIdx(e.Index), e.Name, e.Payload); err != nil {
return err return err
} }
conn.InsertAutocompleteValue(session.SessionID, session.ProjectID, "CUSTOM_Mobile", e.Name) conn.InsertAutocompleteValue(session.SessionID, session.ProjectID, "CUSTOM_MOBILE", e.Name)
return nil return nil
} }
func (conn *Conn) InsertMobileNetworkCall(sess *sessions.Session, e *messages.MobileNetworkCall) error { func (conn *Conn) InsertMobileNetworkCall(sess *sessions.Session, e *messages.MobileNetworkCall) error {
err := conn.InsertRequest(sess.SessionID, e.Timestamp, truncSqIdx(e.Index), e.URL, e.Duration, e.Status < 400) err := conn.InsertRequest(sess.SessionID, e.Timestamp, truncSqIdx(e.Index), e.URL, e.Duration, e.Status < 400)
if err == nil { if err == nil {
conn.InsertAutocompleteValue(sess.SessionID, sess.ProjectID, "REQUEST_Mobile", url.DiscardURLQuery(e.URL)) conn.InsertAutocompleteValue(sess.SessionID, sess.ProjectID, "REQUEST_MOBILE", url.DiscardURLQuery(e.URL))
} }
return err return err
} }
@ -36,7 +36,7 @@ func (conn *Conn) InsertMobileClickEvent(sess *sessions.Session, clickEvent *mes
); err != nil { ); err != nil {
return err return err
} }
conn.InsertAutocompleteValue(sess.SessionID, sess.ProjectID, "CLICK_Mobile", clickEvent.Label) conn.InsertAutocompleteValue(sess.SessionID, sess.ProjectID, "CLICK_MOBILE", clickEvent.Label)
return nil return nil
} }
@ -51,7 +51,7 @@ func (conn *Conn) InsertMobileSwipeEvent(sess *sessions.Session, swipeEvent *mes
); err != nil { ); err != nil {
return err return err
} }
conn.InsertAutocompleteValue(sess.SessionID, sess.ProjectID, "SWIPE_Mobile", swipeEvent.Label) conn.InsertAutocompleteValue(sess.SessionID, sess.ProjectID, "SWIPE_MOBILE", swipeEvent.Label)
return nil return nil
} }
@ -66,7 +66,7 @@ func (conn *Conn) InsertMobileInputEvent(sess *sessions.Session, inputEvent *mes
); err != nil { ); err != nil {
return err return err
} }
conn.InsertAutocompleteValue(sess.SessionID, sess.ProjectID, "INPUT_Mobile", inputEvent.Label) conn.InsertAutocompleteValue(sess.SessionID, sess.ProjectID, "INPUT_MOBILE", inputEvent.Label)
return nil return nil
} }

View file

@ -29,6 +29,11 @@ def _get_current_auth_context(request: Request, jwt_payload: dict) -> schemas.Cu
return request.state.currentContext return request.state.currentContext
def _allow_access_to_endpoint(request: Request, current_context: schemas.CurrentContext) -> bool:
return not current_context.service_account \
or request.url.path not in ["/logout", "/api/logout", "/refresh", "/api/refresh"]
class JWTAuth(HTTPBearer): class JWTAuth(HTTPBearer):
def __init__(self, auto_error: bool = True): def __init__(self, auto_error: bool = True):
super(JWTAuth, self).__init__(auto_error=auto_error) super(JWTAuth, self).__init__(auto_error=auto_error)
@ -68,7 +73,10 @@ class JWTAuth(HTTPBearer):
or old_jwt_payload.get("userId") != jwt_payload.get("userId"): or old_jwt_payload.get("userId") != jwt_payload.get("userId"):
raise HTTPException(status_code=status.HTTP_403_FORBIDDEN, detail="Invalid token or expired token.") raise HTTPException(status_code=status.HTTP_403_FORBIDDEN, detail="Invalid token or expired token.")
return _get_current_auth_context(request=request, jwt_payload=jwt_payload) ctx = _get_current_auth_context(request=request, jwt_payload=jwt_payload)
if not _allow_access_to_endpoint(request=request, current_context=ctx):
raise HTTPException(status_code=status.HTTP_403_FORBIDDEN, detail="Unauthorized endpoint.")
return ctx
else: else:
credentials: HTTPAuthorizationCredentials = await super(JWTAuth, self).__call__(request) credentials: HTTPAuthorizationCredentials = await super(JWTAuth, self).__call__(request)
@ -95,7 +103,10 @@ class JWTAuth(HTTPBearer):
raise HTTPException(status_code=status.HTTP_403_FORBIDDEN, detail="Invalid token or expired token.") raise HTTPException(status_code=status.HTTP_403_FORBIDDEN, detail="Invalid token or expired token.")
return _get_current_auth_context(request=request, jwt_payload=jwt_payload) ctx = _get_current_auth_context(request=request, jwt_payload=jwt_payload)
if not _allow_access_to_endpoint(request=request, current_context=ctx):
raise HTTPException(status_code=status.HTTP_403_FORBIDDEN, detail="Unauthorized endpoint.")
return ctx
logger.warning("Invalid authorization code.") logger.warning("Invalid authorization code.")
raise HTTPException(status_code=status.HTTP_400_BAD_REQUEST, detail="Invalid authorization code.") raise HTTPException(status_code=status.HTTP_400_BAD_REQUEST, detail="Invalid authorization code.")

View file

@ -339,10 +339,13 @@ def create_card(project_id, user_id, data: schemas.CardSchema, dashboard=False):
session_data = None session_data = None
if data.metric_type == schemas.MetricType.heat_map: if data.metric_type == schemas.MetricType.heat_map:
if data.session_id is not None: if data.session_id is not None:
session_data = json.dumps({"sessionId": data.session_id}) session_data = {"sessionId": data.session_id}
else: else:
session_data = __get_heat_map_chart(project_id=project_id, user_id=user_id, session_data = __get_heat_map_chart(project_id=project_id, user_id=user_id,
data=data, include_mobs=False) data=data, include_mobs=False)
if session_data is not None:
session_data = {"sessionId": session_data["sessionId"]}
if session_data is not None: if session_data is not None:
# for EE only # for EE only
keys = sessions_mobs. \ keys = sessions_mobs. \
@ -356,8 +359,8 @@ def create_card(project_id, user_id, data: schemas.CardSchema, dashboard=False):
except Exception as e: except Exception as e:
logger.warning(f"!!!Error while tagging: {k} to {tag} for heatMap") logger.warning(f"!!!Error while tagging: {k} to {tag} for heatMap")
logger.error(str(e)) logger.error(str(e))
session_data = json.dumps(session_data)
_data = {"session_data": session_data} _data = {"session_data": json.dumps(session_data) if session_data is not None else None}
for i, s in enumerate(data.series): for i, s in enumerate(data.series):
for k in s.model_dump().keys(): for k in s.model_dump().keys():
_data[f"{k}_{i}"] = s.__getattribute__(k) _data[f"{k}_{i}"] = s.__getattribute__(k)

View file

@ -57,16 +57,16 @@ def get_by_url(project_id, data: schemas.GetHeatMapPayloadSchema):
# f.value, value_key=f_k)) # f.value, value_key=f_k))
if data.click_rage and not has_click_rage_filter: if data.click_rage and not has_click_rage_filter:
constraints.append("""(issues.session_id IS NULL constraints.append("""(issues_t.session_id IS NULL
OR (issues.datetime >= toDateTime(%(startDate)s/1000) OR (issues_t.datetime >= toDateTime(%(startDate)s/1000)
AND issues.datetime <= toDateTime(%(endDate)s/1000) AND issues_t.datetime <= toDateTime(%(endDate)s/1000)
AND issues.project_id = toUInt16(%(project_id)s) AND issues_t.project_id = toUInt16(%(project_id)s)
AND issues.event_type = 'ISSUE' AND issues_t.event_type = 'ISSUE'
AND issues.project_id = toUInt16(%(project_id)s AND issues_t.project_id = toUInt16(%(project_id)s)
AND mis.project_id = toUInt16(%(project_id)s AND mis.project_id = toUInt16(%(project_id)s)
AND mis.type='click_rage'))))""") AND mis.type='click_rage'))""")
query_from += """ LEFT JOIN experimental.events AS issues ON (main_events.session_id=issues.session_id) query_from += """ LEFT JOIN experimental.events AS issues_t ON (main_events.session_id=issues_t.session_id)
LEFT JOIN experimental.issues AS mis ON (issues.issue_id=mis.issue_id)""" LEFT JOIN experimental.issues AS mis ON (issues_t.issue_id=mis.issue_id)"""
with ch_client.ClickHouseClient() as cur: with ch_client.ClickHouseClient() as cur:
query = cur.format(f"""SELECT main_events.normalized_x AS normalized_x, query = cur.format(f"""SELECT main_events.normalized_x AS normalized_x,
main_events.normalized_y AS normalized_y main_events.normalized_y AS normalized_y

View file

@ -442,7 +442,8 @@ def search2_table(data: schemas.SessionsSearchPayloadSchema, project_id: int, de
if metric_format == schemas.MetricExtendedFormatType.session_count: if metric_format == schemas.MetricExtendedFormatType.session_count:
main_query = f"""SELECT COUNT(DISTINCT {main_col}) OVER () AS main_count, main_query = f"""SELECT COUNT(DISTINCT {main_col}) OVER () AS main_count,
{main_col} AS name, {main_col} AS name,
count(DISTINCT session_id) AS session_count count(DISTINCT session_id) AS session_count,
COALESCE(SUM(count(DISTINCT session_id)) OVER (), 0) AS total_sessions
FROM (SELECT s.session_id AS session_id, FROM (SELECT s.session_id AS session_id,
{extra_col} {extra_col}
{query_part}) AS filtred_sessions {query_part}) AS filtred_sessions
@ -470,11 +471,14 @@ def search2_table(data: schemas.SessionsSearchPayloadSchema, project_id: int, de
logging.debug("--------------------") logging.debug("--------------------")
sessions = cur.execute(main_query) sessions = cur.execute(main_query)
count = 0 count = 0
total_sessions = 0
if len(sessions) > 0: if len(sessions) > 0:
count = sessions[0]["main_count"] count = sessions[0]["main_count"]
total_sessions = sessions[0]["total_sessions"]
for s in sessions: for s in sessions:
s.pop("main_count") s.pop("main_count")
sessions = {"count": count, "values": helper.list_to_camel_case(sessions)} s.pop("total_sessions")
sessions = {"total": count, "count": total_sessions, "values": helper.list_to_camel_case(sessions)}
return sessions return sessions
@ -520,7 +524,7 @@ def search_table_of_individual_issues(data: schemas.SessionsSearchPayloadSchema,
total_sessions = 0 total_sessions = 0
issues_count = 0 issues_count = 0
return {"count": issues_count, "totalSessions": total_sessions, "values": issues} return {"total": issues_count, "count": total_sessions, "values": issues}
def __is_valid_event(is_any: bool, event: schemas.SessionSearchEventSchema2): def __is_valid_event(is_any: bool, event: schemas.SessionSearchEventSchema2):
@ -563,7 +567,7 @@ def __get_event_type(event_type: Union[schemas.EventType, schemas.PerformanceEve
schemas.PerformanceEventType.fetch_failed: "REQUEST", schemas.PerformanceEventType.fetch_failed: "REQUEST",
schemas.EventType.error: "CRASH", schemas.EventType.error: "CRASH",
} }
if platform == "ios" and event_type in defs_mobile: if platform != "web" and event_type in defs_mobile:
return defs_mobile.get(event_type) return defs_mobile.get(event_type)
if event_type not in defs: if event_type not in defs:
raise Exception(f"unsupported EventType:{event_type}") raise Exception(f"unsupported EventType:{event_type}")
@ -964,7 +968,7 @@ def search_query_parts_ch(data: schemas.SessionsSearchPayloadSchema, error_statu
value_key=f"custom{i}")) value_key=f"custom{i}"))
full_args = {**full_args, **_multiple_values(event.source, value_key=f"custom{i}")} full_args = {**full_args, **_multiple_values(event.source, value_key=f"custom{i}")}
else: else:
_column = events.EventType.INPUT_IOS.column _column = events.EventType.INPUT_MOBILE.column
event_where.append(f"main.event_type='{__get_event_type(event_type, platform=platform)}'") event_where.append(f"main.event_type='{__get_event_type(event_type, platform=platform)}'")
events_conditions.append({"type": event_where[-1]}) events_conditions.append({"type": event_where[-1]})
if not is_any: if not is_any:
@ -997,7 +1001,7 @@ def search_query_parts_ch(data: schemas.SessionsSearchPayloadSchema, error_statu
event.value, value_key=e_k)) event.value, value_key=e_k))
events_conditions[-1]["condition"] = event_where[-1] events_conditions[-1]["condition"] = event_where[-1]
else: else:
_column = events.EventType.VIEW_IOS.column _column = events.EventType.VIEW_MOBILE.column
event_where.append(f"main.event_type='{__get_event_type(event_type, platform=platform)}'") event_where.append(f"main.event_type='{__get_event_type(event_type, platform=platform)}'")
events_conditions.append({"type": event_where[-1]}) events_conditions.append({"type": event_where[-1]})
if not is_any: if not is_any:
@ -1089,6 +1093,114 @@ def search_query_parts_ch(data: schemas.SessionsSearchPayloadSchema, error_statu
events_conditions[-1]["condition"] = " AND ".join(events_conditions[-1]["condition"]) events_conditions[-1]["condition"] = " AND ".join(events_conditions[-1]["condition"])
# ----- Mobile
elif event_type == events.EventType.CLICK_MOBILE.ui_type:
_column = events.EventType.CLICK_MOBILE.column
event_where.append(f"main.event_type='{__get_event_type(event_type, platform=platform)}'")
events_conditions.append({"type": event_where[-1]})
if not is_any:
if is_not:
event_where.append(_multiple_conditions(f"sub.{_column} {op} %({e_k})s", event.value,
value_key=e_k))
events_conditions_not.append(
{"type": f"sub.event_type='{__get_event_type(event_type, platform=platform)}'"})
events_conditions_not[-1]["condition"] = event_where[-1]
else:
event_where.append(_multiple_conditions(f"main.{_column} {op} %({e_k})s", event.value,
value_key=e_k))
events_conditions[-1]["condition"] = event_where[-1]
elif event_type == events.EventType.INPUT_MOBILE.ui_type:
_column = events.EventType.INPUT_MOBILE.column
event_where.append(f"main.event_type='{__get_event_type(event_type, platform=platform)}'")
events_conditions.append({"type": event_where[-1]})
if not is_any:
if is_not:
event_where.append(_multiple_conditions(f"sub.{_column} {op} %({e_k})s", event.value,
value_key=e_k))
events_conditions_not.append(
{"type": f"sub.event_type='{__get_event_type(event_type, platform=platform)}'"})
events_conditions_not[-1]["condition"] = event_where[-1]
else:
event_where.append(_multiple_conditions(f"main.{_column} {op} %({e_k})s", event.value,
value_key=e_k))
events_conditions[-1]["condition"] = event_where[-1]
elif event_type == events.EventType.VIEW_MOBILE.ui_type:
_column = events.EventType.VIEW_MOBILE.column
event_where.append(f"main.event_type='{__get_event_type(event_type, platform=platform)}'")
events_conditions.append({"type": event_where[-1]})
if not is_any:
if is_not:
event_where.append(_multiple_conditions(f"sub.{_column} {op} %({e_k})s", event.value,
value_key=e_k))
events_conditions_not.append(
{"type": f"sub.event_type='{__get_event_type(event_type, platform=platform)}'"})
events_conditions_not[-1]["condition"] = event_where[-1]
else:
event_where.append(_multiple_conditions(f"main.{_column} {op} %({e_k})s",
event.value, value_key=e_k))
events_conditions[-1]["condition"] = event_where[-1]
elif event_type == events.EventType.CUSTOM_MOBILE.ui_type:
_column = events.EventType.CUSTOM_MOBILE.column
event_where.append(f"main.event_type='{__get_event_type(event_type, platform=platform)}'")
events_conditions.append({"type": event_where[-1]})
if not is_any:
if is_not:
event_where.append(_multiple_conditions(f"sub.{_column} {op} %({e_k})s", event.value,
value_key=e_k))
events_conditions_not.append(
{"type": f"sub.event_type='{__get_event_type(event_type, platform=platform)}'"})
events_conditions_not[-1]["condition"] = event_where[-1]
else:
event_where.append(_multiple_conditions(f"main.{_column} {op} %({e_k})s",
event.value, value_key=e_k))
events_conditions[-1]["condition"] = event_where[-1]
elif event_type == events.EventType.REQUEST_MOBILE.ui_type:
event_from = event_from % f"{MAIN_EVENTS_TABLE} AS main "
_column = 'url_path'
event_where.append(f"main.event_type='{__get_event_type(event_type, platform=platform)}'")
events_conditions.append({"type": event_where[-1]})
if not is_any:
if is_not:
event_where.append(_multiple_conditions(f"sub.{_column} {op} %({e_k})s", event.value,
value_key=e_k))
events_conditions_not.append(
{"type": f"sub.event_type='{__get_event_type(event_type, platform=platform)}'"})
events_conditions_not[-1]["condition"] = event_where[-1]
else:
event_where.append(_multiple_conditions(f"main.{_column} {op} %({e_k})s", event.value,
value_key=e_k))
events_conditions[-1]["condition"] = event_where[-1]
elif event_type == events.EventType.CRASH_MOBILE.ui_type:
_column = events.EventType.CRASH_MOBILE.column
event_where.append(f"main.event_type='{__get_event_type(event_type, platform=platform)}'")
events_conditions.append({"type": event_where[-1]})
if not is_any:
if is_not:
event_where.append(_multiple_conditions(f"sub.{_column} {op} %({e_k})s", event.value,
value_key=e_k))
events_conditions_not.append(
{"type": f"sub.event_type='{__get_event_type(event_type, platform=platform)}'"})
events_conditions_not[-1]["condition"] = event_where[-1]
else:
event_where.append(_multiple_conditions(f"main.{_column} {op} %({e_k})s",
event.value, value_key=e_k))
events_conditions[-1]["condition"] = event_where[-1]
elif event_type == events.EventType.SWIPE_MOBILE.ui_type and platform != "web":
_column = events.EventType.SWIPE_MOBILE.column
event_where.append(f"main.event_type='{__get_event_type(event_type, platform=platform)}'")
events_conditions.append({"type": event_where[-1]})
if not is_any:
if is_not:
event_where.append(_multiple_conditions(f"sub.{_column} {op} %({e_k})s", event.value,
value_key=e_k))
events_conditions_not.append(
{"type": f"sub.event_type='{__get_event_type(event_type, platform=platform)}'"})
events_conditions_not[-1]["condition"] = event_where[-1]
else:
event_where.append(_multiple_conditions(f"main.{_column} {op} %({e_k})s",
event.value, value_key=e_k))
events_conditions[-1]["condition"] = event_where[-1]
elif event_type == schemas.PerformanceEventType.fetch_failed: elif event_type == schemas.PerformanceEventType.fetch_failed:
event_from = event_from % f"{MAIN_EVENTS_TABLE} AS main " event_from = event_from % f"{MAIN_EVENTS_TABLE} AS main "
_column = 'url_path' _column = 'url_path'

View file

@ -185,8 +185,9 @@ def __filter_subquery(project_id: int, filters: Optional[schemas.SessionsSearchP
errors_only=True, favorite_only=None, errors_only=True, favorite_only=None,
issue=None, user_id=None) issue=None, user_id=None)
params = {**params, **qp_params} params = {**params, **qp_params}
# TODO: test if this line impacts other cards beside insights # This line was added because insights is failing when you add filter steps,
# sub_query = f"INNER JOIN {sub_query} USING(session_id)" # for example when you add a LOCATION filter
sub_query = f"INNER JOIN {sub_query} USING(session_id)"
return params, sub_query return params, sub_query

View file

@ -397,12 +397,19 @@ func (c *connectorImpl) InsertWebClickEvent(session *sessions.Session, msg *mess
if msg.Label == "" { if msg.Label == "" {
return nil return nil
} }
var nX *uint8 = nil var nX *float32 = nil
var nY *uint8 = nil var nY *float32 = nil
if msg.NormalizedX <= 100 && msg.NormalizedY <= 100 { if msg.NormalizedX != 101 && msg.NormalizedY != 101 {
nXVal := uint8(msg.NormalizedX) // To support previous versions of tracker
if msg.NormalizedX <= 100 && msg.NormalizedY <= 100 {
msg.NormalizedX *= 100
msg.NormalizedY *= 100
}
normalizedX := float32(msg.NormalizedX) / 100.0
normalizedY := float32(msg.NormalizedY) / 100.0
nXVal := normalizedX
nX = &nXVal nX = &nXVal
nYVal := uint8(msg.NormalizedY) nYVal := normalizedY
nY = &nYVal nY = &nYVal
} }
if err := c.batches["clicks"].Append( if err := c.batches["clicks"].Append(

View file

@ -3,8 +3,8 @@ CREATE OR REPLACE FUNCTION openreplay_version AS() -> 'v1.19.0-ee';
DROP TABLE IF EXISTS experimental.events_l7d_mv; DROP TABLE IF EXISTS experimental.events_l7d_mv;
ALTER TABLE experimental.events ALTER TABLE experimental.events
ADD COLUMN IF NOT EXISTS normalized_x Nullable(UInt8), ADD COLUMN IF NOT EXISTS normalized_x Nullable(Float32),
ADD COLUMN IF NOT EXISTS normalized_y Nullable(UInt8), ADD COLUMN IF NOT EXISTS normalized_y Nullable(Float32),
DROP COLUMN IF EXISTS coordinate; DROP COLUMN IF EXISTS coordinate;
CREATE MATERIALIZED VIEW IF NOT EXISTS experimental.events_l7d_mv CREATE MATERIALIZED VIEW IF NOT EXISTS experimental.events_l7d_mv

View file

@ -81,8 +81,8 @@ CREATE TABLE IF NOT EXISTS experimental.events
error_tags_values Array(Nullable(String)), error_tags_values Array(Nullable(String)),
transfer_size Nullable(UInt32), transfer_size Nullable(UInt32),
selector Nullable(String), selector Nullable(String),
normalized_x Nullable(UInt8), normalized_x Nullable(Float32),
normalized_y Nullable(UInt8), normalized_y Nullable(Float32),
message_id UInt64 DEFAULT 0, message_id UInt64 DEFAULT 0,
_timestamp DateTime DEFAULT now() _timestamp DateTime DEFAULT now()
) ENGINE = ReplacingMergeTree(_timestamp) ) ENGINE = ReplacingMergeTree(_timestamp)

View file

@ -19,8 +19,8 @@ $fn_def$, :'next_version')
-- --
ALTER TABLE IF EXISTS events.clicks ALTER TABLE IF EXISTS events.clicks
ADD COLUMN IF NOT EXISTS normalized_x smallint NULL, ADD COLUMN IF NOT EXISTS normalized_x decimal NULL,
ADD COLUMN IF NOT EXISTS normalized_y smallint NULL, ADD COLUMN IF NOT EXISTS normalized_y decimal NULL,
DROP COLUMN IF EXISTS x, DROP COLUMN IF EXISTS x,
DROP COLUMN IF EXISTS y; DROP COLUMN IF EXISTS y;

View file

@ -659,16 +659,16 @@ CREATE INDEX pages_query_nn_gin_idx ON events.pages USING GIN (query gin_trgm_op
CREATE TABLE events.clicks CREATE TABLE events.clicks
( (
session_id bigint NOT NULL REFERENCES public.sessions (session_id) ON DELETE CASCADE, session_id bigint NOT NULL REFERENCES public.sessions (session_id) ON DELETE CASCADE,
message_id bigint NOT NULL, message_id bigint NOT NULL,
timestamp bigint NOT NULL, timestamp bigint NOT NULL,
label text DEFAULT NULL, label text DEFAULT NULL,
url text DEFAULT '' NOT NULL, url text DEFAULT '' NOT NULL,
path text, path text,
selector text DEFAULT '' NOT NULL, selector text DEFAULT '' NOT NULL,
hesitation integer DEFAULT NULL, hesitation integer DEFAULT NULL,
normalized_x smallint DEFAULT NULL, normalized_x decimal DEFAULT NULL,
normalized_y smallint DEFAULT NULL, normalized_y decimal DEFAULT NULL,
PRIMARY KEY (session_id, message_id) PRIMARY KEY (session_id, message_id)
); );
CREATE INDEX clicks_session_id_idx ON events.clicks (session_id); CREATE INDEX clicks_session_id_idx ON events.clicks (session_id);

View file

@ -62,9 +62,9 @@ function ClickMapCard({
if (mapUrl) return evt.path.includes(mapUrl) if (mapUrl) return evt.path.includes(mapUrl)
return evt return evt
}) || { timestamp: metricStore.instance.data.startTs } }) || { timestamp: metricStore.instance.data.startTs }
const ts = jumpToEvent.timestamp ?? metricStore.instance.data.startTs
const jumpTimestamp = (jumpToEvent.timestamp - metricStore.instance.data.startTs) + jumpToEvent.domBuildingTime + 99 // 99ms safety margin to give some time for the DOM to load const domTime = jumpToEvent.domBuildingTime ?? 0
const jumpTimestamp = (ts - metricStore.instance.data.startTs) + domTime + 99 // 99ms safety margin to give some time for the DOM to load
return ( return (
<div id="clickmap-render"> <div id="clickmap-render">
<ClickMapRenderer <ClickMapRenderer

View file

@ -51,8 +51,8 @@ function WebPlayer(props: any) {
const isPlayerReady = contextValue.store?.get().ready const isPlayerReady = contextValue.store?.get().ready
React.useEffect(() => { React.useEffect(() => {
contextValue.player && contextValue.player.play() contextValue.player && contextValue.player.play()
if (isPlayerReady && insights.size > 0) { if (isPlayerReady && insights.size > 0 && jumpTimestamp) {
setTimeout(() => { setTimeout(() => {
contextValue.player.pause() contextValue.player.pause()
contextValue.player.jump(jumpTimestamp) contextValue.player.jump(jumpTimestamp)

View file

@ -9,7 +9,7 @@
height: 100%; height: 100%;
/* border: solid thin $gray-light; */ /* border: solid thin $gray-light; */
/* border-radius: 3px; */ /* border-radius: 3px; */
overflow: hidden; overflow-y: scroll;
} }
.checkers { .checkers {

View file

@ -302,8 +302,8 @@ export default class Widget {
} else if (this.metricType === FUNNEL) { } else if (this.metricType === FUNNEL) {
_data.funnel = new Funnel().fromJSON(_data); _data.funnel = new Funnel().fromJSON(_data);
} else if (this.metricType === TABLE) { } else if (this.metricType === TABLE) {
const totalSessions = data[0]['totalSessions']; const count = data[0]['count'];
_data[0]['values'] = data[0]['values'].map((s: any) => new SessionsByRow().fromJson(s, totalSessions, this.metricOf)); _data[0]['values'] = data[0]['values'].map((s: any) => new SessionsByRow().fromJson(s, count, this.metricOf));
} else { } else {
if (data.hasOwnProperty('chart')) { if (data.hasOwnProperty('chart')) {
_data['value'] = data.value; _data['value'] = data.value;

View file

@ -78,10 +78,10 @@ export interface State extends ScreenState, ListsState {
} }
const userEvents = [ const userEvents = [
MType.IosSwipeEvent, MType.MobileSwipeEvent,
MType.IosClickEvent, MType.MobileClickEvent,
MType.IosInputEvent, MType.MobileInputEvent,
MType.IosScreenChanges, MType.MobileScreenChanges,
]; ];
export default class IOSMessageManager implements IMessageManager { export default class IOSMessageManager implements IMessageManager {
@ -233,7 +233,7 @@ export default class IOSMessageManager implements IMessageManager {
} }
switch (msg.tp) { switch (msg.tp) {
case MType.IosPerformanceEvent: case MType.MobilePerformanceEvent:
const performanceStats = ['background', 'memoryUsage', 'mainThreadCPU']; const performanceStats = ['background', 'memoryUsage', 'mainThreadCPU'];
if (performanceStats.includes(msg.name)) { if (performanceStats.includes(msg.name)) {
this.performanceManager.append(msg); this.performanceManager.append(msg);
@ -253,21 +253,21 @@ export default class IOSMessageManager implements IMessageManager {
// case MType.IosInputEvent: // case MType.IosInputEvent:
// console.log('input', msg) // console.log('input', msg)
// break; // break;
case MType.IosNetworkCall: case MType.MobileNetworkCall:
this.lists.lists.fetch.insert(getResourceFromNetworkRequest(msg, this.sessionStart)); this.lists.lists.fetch.insert(getResourceFromNetworkRequest(msg, this.sessionStart));
break; break;
case MType.WsChannel: case MType.WsChannel:
this.lists.lists.websocket.insert(msg); this.lists.lists.websocket.insert(msg);
break; break;
case MType.IosEvent: case MType.MobileEvent:
// @ts-ignore // @ts-ignore
this.lists.lists.event.insert({ ...msg, source: 'openreplay' }); this.lists.lists.event.insert({ ...msg, source: 'openreplay' });
break; break;
case MType.IosSwipeEvent: case MType.MobileSwipeEvent:
case MType.IosClickEvent: case MType.MobileClickEvent:
this.touchManager.append(msg); this.touchManager.append(msg);
break; break;
case MType.IosLog: case MType.MobileLog:
const log = { ...msg, level: msg.severity }; const log = { ...msg, level: msg.severity };
// @ts-ignore // @ts-ignore
this.lists.lists.log.append(Log(log)); this.lists.lists.log.append(Log(log));

View file

@ -31,7 +31,7 @@ export default class TouchManager extends ListWalker<IosClickEvent | IosSwipeEve
public move(t: number) { public move(t: number) {
const lastTouch = this.moveGetLast(t) const lastTouch = this.moveGetLast(t)
if (!!lastTouch) { if (!!lastTouch) {
if (lastTouch.tp === MType.IosSwipeEvent) { if (lastTouch.tp === MType.MobileSwipeEvent) {
return return
// not using swipe rn // not using swipe rn
// this.touchTrail?.createSwipeTrail({ // this.touchTrail?.createSwipeTrail({

View file

@ -233,10 +233,10 @@ export default class Screen {
break; break;
case ScaleMode.AdjustParentHeight: case ScaleMode.AdjustParentHeight:
// we want to scale the document with true height so the clickmap will be scrollable // we want to scale the document with true height so the clickmap will be scrollable
const usedHeight = const usedHeight = height + 'px';
this.document?.body.scrollHeight && this.document?.body.scrollHeight > height // this.document?.body.scrollHeight && this.document?.body.scrollHeight > height
? this.document.body.scrollHeight + 'px' // ? this.document.body.scrollHeight + 'px'
: height + 'px'; // : height + 'px';
this.scaleRatio = offsetWidth / width; this.scaleRatio = offsetWidth / width;
translate = 'translate(-50%, 0)'; translate = 'translate(-50%, 0)';
posStyles = { top: 0, height: usedHeight }; posStyles = { top: 0, height: usedHeight };

View file

@ -146,39 +146,37 @@ export default class TargetMarker {
if (clicks && this.screen.document) { if (clicks && this.screen.document) {
this.clickMapOverlay?.remove(); this.clickMapOverlay?.remove();
const overlay = document.createElement('canvas'); const overlay = document.createElement('canvas');
const iframeSize = this.screen.iframeStylesRef;
const scrollHeight = this.screen.document?.documentElement.scrollHeight || 0; const scrollHeight = this.screen.document?.documentElement.scrollHeight || 0;
const scrollWidth = this.screen.document?.documentElement.scrollWidth || 0; const scrollWidth = this.screen.document?.documentElement.scrollWidth || 0;
const scaleRatio = this.screen.getScale();
Object.assign( Object.assign(
overlay.style, overlay.style,
clickmapStyles.overlayStyle({ clickmapStyles.overlayStyle({
height: iframeSize.height, height: scrollHeight + 'px',
width: iframeSize.width, width: scrollWidth + 'px',
scale: scaleRatio,
}) })
); );
this.clickMapOverlay = overlay; this.clickMapOverlay = overlay;
this.screen.getParentElement()?.appendChild(overlay); this.screen.document.body.appendChild(overlay);
const pointMap: Record<string, { times: number; data: number[], original: any }> = {}; const pointMap: Record<string, { times: number; data: number[], original: any }> = {};
const ovWidth = parseInt(iframeSize.width); overlay.width = scrollWidth;
const ovHeight = parseInt(iframeSize.height); overlay.height = scrollHeight;
overlay.width = ovWidth;
overlay.height = ovHeight;
let maxIntensity = 0; let maxIntensity = 0;
clicks.forEach((point) => { clicks.forEach((point) => {
const key = `${point.normalizedY}-${point.normalizedX}`; const y = roundToSecond(point.normalizedY);
const x = roundToSecond(point.normalizedX);
const key = `${y}-${x}`;
if (pointMap[key]) { if (pointMap[key]) {
const times = pointMap[key].times + 1; const times = pointMap[key].times + 1;
maxIntensity = Math.max(maxIntensity, times); maxIntensity = Math.max(maxIntensity, times);
pointMap[key].times = times; pointMap[key].times = times;
} else { } else {
const clickData = [ const clickData = [
(point.normalizedX / 100) * scrollWidth, (x / 100) * scrollWidth,
(point.normalizedY / 100) * scrollHeight, (y / 100) * scrollHeight,
]; ];
pointMap[key] = { times: 1, data: clickData, original: point }; pointMap[key] = { times: 1, data: clickData, original: point };
} }
@ -204,3 +202,7 @@ export default class TargetMarker {
} }
} }
} }
function roundToSecond(num: number) {
return Math.round(num * 100) / 100;
}

View file

@ -1,14 +1,12 @@
export const clickmapStyles = { export const clickmapStyles = {
overlayStyle: ({ height, width, scale }: { height: string, width: string, scale: number }) => ({ overlayStyle: ({ height, width }: { height: string, width: string }) => ({
transform: `scale(${scale}) translate(-50%, 0)`,
position: 'absolute', position: 'absolute',
top: '0px', top: '0px',
left: '50%', left: 0,
width, width,
height, height,
background: 'rgba(0,0,0, 0.15)', background: 'rgba(0,0,0, 0.15)',
zIndex: 9 * 10e3, zIndex: 9 * 10e3,
transformOrigin: 'left top',
}), }),
totalClicks: { totalClicks: {
fontSize: '16px', fontSize: '16px',

View file

@ -1,7 +1,6 @@
apiVersion: v2 apiVersion: v2
name: chalice name: chalice
description: A Helm chart for Kubernetes description: A Helm chart for Kubernetes
# A chart can be either an 'application' or a 'library' chart. # A chart can be either an 'application' or a 'library' chart.
# #
# Application charts are a collection of templates that can be packaged into versioned archives # Application charts are a collection of templates that can be packaged into versioned archives
@ -11,14 +10,12 @@ description: A Helm chart for Kubernetes
# a dependency of application charts to inject those utilities and functions into the rendering # a dependency of application charts to inject those utilities and functions into the rendering
# pipeline. Library charts do not define any templates and therefore cannot be deployed. # pipeline. Library charts do not define any templates and therefore cannot be deployed.
type: application type: application
# This is the chart version. This version number should be incremented each time you make changes # This is the chart version. This version number should be incremented each time you make changes
# to the chart and its templates, including the app version. # to the chart and its templates, including the app version.
# Versions are expected to follow Semantic Versioning (https://semver.org/) # Versions are expected to follow Semantic Versioning (https://semver.org/)
version: 0.1.7 version: 0.1.7
# This is the version number of the application being deployed. This version number should be # This is the version number of the application being deployed. This version number should be
# incremented each time you make changes to the application. Versions are not expected to # incremented each time you make changes to the application. Versions are not expected to
# follow Semantic Versioning. They should reflect the version the application is using. # follow Semantic Versioning. They should reflect the version the application is using.
# It is recommended to use it with quotes. # It is recommended to use it with quotes.
AppVersion: "v1.19.0" AppVersion: "v1.19.6"

View file

@ -1,7 +1,6 @@
apiVersion: v2 apiVersion: v2
name: db name: db
description: A Helm chart for Kubernetes description: A Helm chart for Kubernetes
# A chart can be either an 'application' or a 'library' chart. # A chart can be either an 'application' or a 'library' chart.
# #
# Application charts are a collection of templates that can be packaged into versioned archives # Application charts are a collection of templates that can be packaged into versioned archives
@ -11,14 +10,12 @@ description: A Helm chart for Kubernetes
# a dependency of application charts to inject those utilities and functions into the rendering # a dependency of application charts to inject those utilities and functions into the rendering
# pipeline. Library charts do not define any templates and therefore cannot be deployed. # pipeline. Library charts do not define any templates and therefore cannot be deployed.
type: application type: application
# This is the chart version. This version number should be incremented each time you make changes # This is the chart version. This version number should be incremented each time you make changes
# to the chart and its templates, including the app version. # to the chart and its templates, including the app version.
# Versions are expected to follow Semantic Versioning (https://semver.org/) # Versions are expected to follow Semantic Versioning (https://semver.org/)
version: 0.1.1 version: 0.1.1
# This is the version number of the application being deployed. This version number should be # This is the version number of the application being deployed. This version number should be
# incremented each time you make changes to the application. Versions are not expected to # incremented each time you make changes to the application. Versions are not expected to
# follow Semantic Versioning. They should reflect the version the application is using. # follow Semantic Versioning. They should reflect the version the application is using.
# It is recommended to use it with quotes. # It is recommended to use it with quotes.
AppVersion: "v1.19.0" AppVersion: "v1.19.1"

View file

@ -1,7 +1,6 @@
apiVersion: v2 apiVersion: v2
name: frontend name: frontend
description: A Helm chart for Kubernetes description: A Helm chart for Kubernetes
# A chart can be either an 'application' or a 'library' chart. # A chart can be either an 'application' or a 'library' chart.
# #
# Application charts are a collection of templates that can be packaged into versioned archives # Application charts are a collection of templates that can be packaged into versioned archives
@ -11,14 +10,12 @@ description: A Helm chart for Kubernetes
# a dependency of application charts to inject those utilities and functions into the rendering # a dependency of application charts to inject those utilities and functions into the rendering
# pipeline. Library charts do not define any templates and therefore cannot be deployed. # pipeline. Library charts do not define any templates and therefore cannot be deployed.
type: application type: application
# This is the chart version. This version number should be incremented each time you make changes # This is the chart version. This version number should be incremented each time you make changes
# to the chart and its templates, including the app version. # to the chart and its templates, including the app version.
# Versions are expected to follow Semantic Versioning (frontends://semver.org/) # Versions are expected to follow Semantic Versioning (frontends://semver.org/)
version: 0.1.10 version: 0.1.10
# This is the version number of the application being deployed. This version number should be # This is the version number of the application being deployed. This version number should be
# incremented each time you make changes to the application. Versions are not expected to # incremented each time you make changes to the application. Versions are not expected to
# follow Semantic Versioning. They should reflect the version the application is using. # follow Semantic Versioning. They should reflect the version the application is using.
# It is recommended to use it with quotes. # It is recommended to use it with quotes.
AppVersion: "v1.19.0" AppVersion: "v1.19.2"

View file

@ -8,7 +8,11 @@ RETENTION_TIME=${RETENTION_TIME:-345600000}
topics=( topics=(
"raw" "raw"
"raw-ios" "raw-ios"
"raw-images"
"canvas-images"
"trigger" "trigger"
"canvas-trigger"
"mobile-trigger"
"cache" "cache"
"analytics" "analytics"
"storage-failover" "storage-failover"

View file

@ -19,8 +19,8 @@ $fn_def$, :'next_version')
-- --
ALTER TABLE IF EXISTS events.clicks ALTER TABLE IF EXISTS events.clicks
ADD COLUMN IF NOT EXISTS normalized_x smallint NULL, ADD COLUMN IF NOT EXISTS normalized_x decimal NULL,
ADD COLUMN IF NOT EXISTS normalized_y smallint NULL, ADD COLUMN IF NOT EXISTS normalized_y decimal NULL,
DROP COLUMN IF EXISTS x, DROP COLUMN IF EXISTS x,
DROP COLUMN IF EXISTS y; DROP COLUMN IF EXISTS y;

View file

@ -620,16 +620,16 @@ CREATE INDEX pages_query_nn_gin_idx ON events.pages USING GIN (query gin_trgm_op
CREATE TABLE events.clicks CREATE TABLE events.clicks
( (
session_id bigint NOT NULL REFERENCES public.sessions (session_id) ON DELETE CASCADE, session_id bigint NOT NULL REFERENCES public.sessions (session_id) ON DELETE CASCADE,
message_id bigint NOT NULL, message_id bigint NOT NULL,
timestamp bigint NOT NULL, timestamp bigint NOT NULL,
label text DEFAULT NULL, label text DEFAULT NULL,
url text DEFAULT '' NOT NULL, url text DEFAULT '' NOT NULL,
path text, path text,
selector text DEFAULT '' NOT NULL, selector text DEFAULT '' NOT NULL,
hesitation integer DEFAULT NULL, hesitation integer DEFAULT NULL,
normalized_x smallint DEFAULT NULL, normalized_x decimal DEFAULT NULL,
normalized_y smallint DEFAULT NULL, normalized_y decimal DEFAULT NULL,
PRIMARY KEY (session_id, message_id) PRIMARY KEY (session_id, message_id)
); );
CREATE INDEX clicks_session_id_idx ON events.clicks (session_id); CREATE INDEX clicks_session_id_idx ON events.clicks (session_id);

View file

@ -18,13 +18,13 @@ returns `result` without changes.
```js ```js
import Tracker from '@openreplay/tracker'; import Tracker from '@openreplay/tracker';
import trackerGraphQL from '@openreplay/tracker-graphql'; import { createGraphqlMiddleware } from '@openreplay/tracker-graphql';
const tracker = new Tracker({ const tracker = new Tracker({
projectKey: YOUR_PROJECT_KEY, projectKey: YOUR_PROJECT_KEY,
}); });
export const recordGraphQL = tracker.plugin(trackerGraphQL()); export const recordGraphQL = tracker.use(createGraphqlMiddleware());
``` ```
### Relay ### Relay
@ -33,15 +33,28 @@ If you're using [Relay network tools](https://github.com/relay-tools/react-relay
you can simply [create a middleware](https://github.com/relay-tools/react-relay-network-modern/tree/master?tab=readme-ov-file#example-of-injecting-networklayer-with-middlewares-on-the-client-side) you can simply [create a middleware](https://github.com/relay-tools/react-relay-network-modern/tree/master?tab=readme-ov-file#example-of-injecting-networklayer-with-middlewares-on-the-client-side)
```js ```js
import { createRelayMiddleware } from '@openreplay/tracker-graphql' import { createRelayMiddleware } from '@openreplay/tracker-graphql';
const trackerMiddleware = createRelayMiddleware(tracker) const trackerMiddleware = tracker.use(createRelayMiddleware());
const network = new RelayNetworkLayer([ const network = new RelayNetworkLayer([
// your middleware // your middleware
// , // ,
trackerMiddleware trackerMiddleware,
]) ]);
```
You can pass a Sanitizer function to `createRelayMiddleware` to sanitize the variables and data before sending them to OpenReplay.
```js
const trackerLink = tracker.use(
createRelayMiddleware((variables) => {
return {
...variables,
password: '***',
};
}),
);
``` ```
Or you can manually put `recordGraphQL` call Or you can manually put `recordGraphQL` call
@ -52,22 +65,22 @@ then you should do something like below
import { createGraphqlMiddleware } from '@openreplay/tracker-graphql'; // see above for recordGraphQL definition import { createGraphqlMiddleware } from '@openreplay/tracker-graphql'; // see above for recordGraphQL definition
import { Environment } from 'relay-runtime'; import { Environment } from 'relay-runtime';
const handler = createGraphqlMiddleware(tracker) const handler = tracker.use(createGraphqlMiddleware());
function fetchQuery(operation, variables, cacheConfig, uploadables) { function fetchQuery(operation, variables, cacheConfig, uploadables) {
return fetch('www.myapi.com/resource', { return fetch('www.myapi.com/resource', {
// ... // ...
}) })
.then(response => response.json()) .then((response) => response.json())
.then(result => .then((result) =>
handler( handler(
// op kind, name, variables, response, duration (default 0) // op kind, name, variables, response, duration (default 0)
operation.operationKind, operation.operationKind,
operation.name, operation.name,
variables, variables,
result, result,
duration, duration,
), ),
); );
} }
@ -81,10 +94,23 @@ See [Relay Network Layer](https://relay.dev/docs/en/network-layer) for details.
For [Apollo](https://www.apollographql.com/) you should create a new `ApolloLink` For [Apollo](https://www.apollographql.com/) you should create a new `ApolloLink`
```js ```js
import { createTrackerLink } from '@openreplay/tracker-graphql' import { createTrackerLink } from '@openreplay/tracker-graphql';
const trackerLink = createTrackerLink(tracker); const trackerLink = tracker.use(createTrackerLink());
const yourLink = new ApolloLink(trackerLink) const yourLink = new ApolloLink(trackerLink);
```
You can pass a Sanitizer function to `createRelayMiddleware` to sanitize the variables and data before sending them to OpenReplay.
```js
const trackerLink = tracker.use(
createTrackerLink((variables) => {
return {
...variables,
password: '***',
};
}),
);
``` ```
Alternatively you can use generic graphql handler: Alternatively you can use generic graphql handler:
@ -93,18 +119,21 @@ Alternatively you can use generic graphql handler:
import { createGraphqlMiddleware } from '@openreplay/tracker-graphql'; // see above for recordGraphQL definition import { createGraphqlMiddleware } from '@openreplay/tracker-graphql'; // see above for recordGraphQL definition
import { ApolloLink } from 'apollo-link'; import { ApolloLink } from 'apollo-link';
const handler = createGraphqlMiddleware(tracker) const handler = tracker.use(createGraphqlMiddleware());
const trackerApolloLink = new ApolloLink((operation, forward) => { const trackerApolloLink = new ApolloLink((operation, forward) => {
return forward(operation).map(result => operation.setContext({ start: performance.now() });
handler( return forward(operation).map((result) => {
const time = performance.now() - operation.getContext().start;
return handler(
// op kind, name, variables, response, duration (default 0) // op kind, name, variables, response, duration (default 0)
operation.query.definitions[0].operation, operation.query.definitions[0].operation,
operation.operationName, operation.operationName,
operation.variables, operation.variables,
result, result,
), time,
); );
});
}); });
const link = ApolloLink.from([ const link = ApolloLink.from([

View file

@ -1,5 +1,6 @@
import { App, Messages } from '@openreplay/tracker'; import { App, Messages } from '@openreplay/tracker';
import Observable from 'zen-observable'; import Observable from 'zen-observable';
import { Sanitizer } from './types';
type Operation = { type Operation = {
query: Record<string, any>; query: Record<string, any>;
@ -9,48 +10,63 @@ type Operation = {
}; };
type NextLink = (operation: Operation) => Observable<Record<string, any>>; type NextLink = (operation: Operation) => Observable<Record<string, any>>;
export const createTrackerLink = (app: App | null) => { export const createTrackerLink = (
if (!app) { sanitizer?: Sanitizer<Record<string, any> | undefined | null>,
return (operation: Operation, forward: NextLink) => forward(operation); ) => {
} return (app: App | null) => {
return (operation: Operation, forward: NextLink) => { if (!app) {
return new Observable((observer) => { return (operation: Operation, forward: NextLink) => forward(operation);
const start = app.timestamp(); }
const observable = forward(operation); return (operation: Operation, forward: NextLink) => {
const subscription = observable.subscribe({ return new Observable((observer) => {
next(value) { const start = app.timestamp();
const end = app.timestamp(); const observable = forward(operation);
app.send( const subscription = observable.subscribe({
Messages.GraphQL( next(value) {
operation.query.definitions[0].kind, const end = app.timestamp();
operation.operationName, const operationDefinition = operation.query.definitions[0];
JSON.stringify(operation.variables), app.send(
JSON.stringify(value.data), Messages.GraphQL(
end - start, operationDefinition.kind === 'OperationDefinition'
), ? operationDefinition.operation
); : 'unknown?',
observer.next(value); operation.operationName,
}, JSON.stringify(
error(error) { sanitizer
const end = app.timestamp(); ? sanitizer(operation.variables)
app.send( : operation.variables,
Messages.GraphQL( ),
operation.query.definitions[0].kind, JSON.stringify(sanitizer ? sanitizer(value.data) : value.data),
operation.operationName, end - start,
JSON.stringify(operation.variables), ),
JSON.stringify(error), );
end - start, observer.next(value);
), },
); error(error) {
observer.error(error); const end = app.timestamp();
}, app.send(
complete() { Messages.GraphQL(
observer.complete(); operation.query.definitions[0].kind,
}, operation.operationName,
}); JSON.stringify(
sanitizer
? sanitizer(operation.variables)
: operation.variables,
),
JSON.stringify(error),
end - start,
),
);
observer.error(error);
},
complete() {
observer.complete();
},
});
return () => subscription.unsubscribe(); return () => subscription.unsubscribe();
}); });
};
}; };
}; };

View file

@ -1,4 +1,4 @@
import { App, Messages } from "@openreplay/tracker"; import { App, Messages } from '@openreplay/tracker';
function createGraphqlMiddleware() { function createGraphqlMiddleware() {
return (app: App | null) => { return (app: App | null) => {
@ -10,7 +10,7 @@ function createGraphqlMiddleware() {
operationName: string, operationName: string,
variables: any, variables: any,
result: any, result: any,
duration = 0 duration = 0,
) => { ) => {
try { try {
app.send( app.send(
@ -30,4 +30,4 @@ function createGraphqlMiddleware() {
}; };
} }
export default createGraphqlMiddleware export default createGraphqlMiddleware;

View file

@ -1,9 +1,11 @@
import createTrackerLink from './apolloMiddleware.js'; import createTrackerLink from './apolloMiddleware.js';
import createRelayMiddleware from './relayMiddleware.js'; import createRelayMiddleware from './relayMiddleware.js';
import createGraphqlMiddleware from './graphqlMiddleware.js'; import createGraphqlMiddleware from './graphqlMiddleware.js';
import { Sanitizer } from './types.js';
export { export {
createTrackerLink, createTrackerLink,
createRelayMiddleware, createRelayMiddleware,
createGraphqlMiddleware, createGraphqlMiddleware,
} Sanitizer,
};

View file

@ -1,37 +1,55 @@
import { App, Messages } from '@openreplay/tracker'; import { App, Messages } from '@openreplay/tracker';
import type { Middleware, RelayRequest } from './relaytypes'; import type { Middleware, RelayRequest } from './relaytypes';
import { Sanitizer } from './types';
const createRelayMiddleware = (app: App | null): Middleware => { const createRelayMiddleware = (sanitizer?: Sanitizer<Record<string, any>>) => {
if (!app) { return (app: App | null): Middleware => {
return (next) => async (req) => await next(req); if (!app) {
} return (next) => async (req) => await next(req);
return (next) => async (req) => {
const start = app.timestamp();
const resp = await next(req)
const end = app.timestamp();
if ('requests' in req) {
req.requests.forEach((request) => {
app.send(getMessage(request, resp.json as Record<string, any>, end - start))
})
} else {
app.send(getMessage(req, resp.json as Record<string, any>, end - start))
} }
return resp; return (next) => async (req) => {
} const start = app.timestamp();
const resp = await next(req);
const end = app.timestamp();
if ('requests' in req) {
req.requests.forEach((request) => {
app.send(
getMessage(
request,
resp.json as Record<string, any>,
end - start,
sanitizer,
),
);
});
} else {
app.send(
getMessage(
req,
resp.json as Record<string, any>,
end - start,
sanitizer,
),
);
}
return resp;
};
};
}; };
function getMessage(request: RelayRequest, json: Record<string, any>, duration: number) { function getMessage(
request: RelayRequest,
json: Record<string, any>,
duration: number,
sanitizer?: Sanitizer<Record<string, any>>,
) {
const opKind = request.operation.kind; const opKind = request.operation.kind;
const opName = request.operation.name; const opName = request.operation.name;
const vars = JSON.stringify(request.variables) const vars = JSON.stringify(
const opResp = JSON.stringify(json) sanitizer ? sanitizer(request.variables) : request.variables,
return Messages.GraphQL( );
opKind, const opResp = JSON.stringify(sanitizer ? sanitizer(json) : json);
opName, return Messages.GraphQL(opKind, opName, vars, opResp, duration);
vars,
opResp,
duration
)
} }
export default createRelayMiddleware export default createRelayMiddleware;

View file

@ -1,4 +1,3 @@
type ConcreteBatch = { type ConcreteBatch = {
kind: 'Batch'; kind: 'Batch';
fragment: any; fragment: any;
@ -9,7 +8,7 @@ type ConcreteBatch = {
text: string | null; text: string | null;
operationKind: string; operationKind: string;
}; };
type Variables = { [name: string]: any }; export type Variables = { [name: string]: any };
interface FetchOpts { interface FetchOpts {
url?: string; url?: string;
method: 'POST' | 'GET'; method: 'POST' | 'GET';
@ -17,7 +16,13 @@ interface FetchOpts {
body: string | FormData; body: string | FormData;
credentials?: 'same-origin' | 'include' | 'omit'; credentials?: 'same-origin' | 'include' | 'omit';
mode?: 'cors' | 'websocket' | 'navigate' | 'no-cors' | 'same-origin'; mode?: 'cors' | 'websocket' | 'navigate' | 'no-cors' | 'same-origin';
cache?: 'default' | 'no-store' | 'reload' | 'no-cache' | 'force-cache' | 'only-if-cached'; cache?:
| 'default'
| 'no-store'
| 'reload'
| 'no-cache'
| 'force-cache'
| 'only-if-cached';
redirect?: 'follow' | 'error' | 'manual'; redirect?: 'follow' | 'error' | 'manual';
signal?: AbortSignal; signal?: AbortSignal;
[name: string]: any; [name: string]: any;

View file

@ -0,0 +1 @@
export type Sanitizer<T> = (values: T) => Partial<T>;

View file

@ -963,8 +963,8 @@ export default class App {
deviceMemory, deviceMemory,
jsHeapSizeLimit, jsHeapSizeLimit,
timezone: getTimezone(), timezone: getTimezone(),
width: window.innerWidth, width: window.screen.width,
height: window.innerHeight, height: window.screen.height,
}), }),
}) })
const { const {
@ -1220,7 +1220,9 @@ export default class App {
timezone: getTimezone(), timezone: getTimezone(),
condition: conditionName, condition: conditionName,
assistOnly: startOpts.assistOnly ?? this.socketMode, assistOnly: startOpts.assistOnly ?? this.socketMode,
}), width: window.screen.width,
height: window.screen.height
}),
}) })
if (r.status !== 200) { if (r.status !== 200) {
const error = await r.text() const error = await r.text()