Changes for page Friday 14 February 2025
Last modified by dennis yoshikawa on 2025/02/14 01:23
From version
1.2


edited by Abdurachman Putra
on 2025/02/13 09:36
on 2025/02/13 09:36
Change comment:
There is no comment for this version
To version
6.2


edited by dennis yoshikawa
on 2025/02/14 01:13
on 2025/02/14 01:13
Change comment:
There is no comment for this version
Summary
Details
- Page properties
-
- Parent
-
... ... @@ -1,1 +1,1 @@ 1 - Main.PTGUE.WebHome1 +Weekly Report.February .WebHome - Author
-
... ... @@ -1,1 +1,1 @@ 1 -XWiki.s urya1 +XWiki.dennis - Content
-
... ... @@ -1,6 +1,186 @@ 1 1 = Weekly Report 14 February 2025 = 2 2 3 3 4 -== Data Engineer == 4 +== (% id="cke_bm_29973S" style="display:none" %) (%%)Data Engineer == 5 5 6 ->What Have You Done?/What Issues You Have?What Next You Will Do? (Optional)What Support You Need? (Optional) 6 +>What Have You Done in This Week? 7 + 8 +((( 9 +{{success}} 10 +What Have You Done? 11 +{{/success}} 12 + 13 +* Airflow-airbyte maintenance 14 + 15 +~1. create task for delete gcs folder table dharmadexa 16 +2. change incremental method in airbyte for tb) 17 + 18 + 19 +* *D2D* 20 + 21 +~1. enhance logic stickness rate (mart) 22 +2. weekly regrup w/product 23 + 24 +* MCN 25 + 26 +~1. troubleshoot on DAG and resource 27 +2. repointing new database for all pipeline existing 28 + 29 +* Screening 30 + 31 +~1. regroup w/PM and DA for validate data and discuss funneling data 32 +2. enhance logic fact table dharmadexa 33 +3. preparation dharmadexa data phase 3 (funneling) 34 +4. testing geo.py to get city from long lat 35 +5. migrasi dataset cleaned_screening_merck.dim_screening_stunting to CH 36 + 37 +{{warning}} 38 +What Issues You Have? 39 +{{/warning}} 40 + 41 +1. xxx 42 +1. xxx 43 +1. xxx 44 + 45 +{{info}} 46 +What Next You Will Do? (Optional) 47 +{{/info}} 48 + 49 +1. xxx 50 +1. xxx 51 +1. xxx 52 + 53 +**What Support You Need? (Optional)** 54 + 55 + 56 +---- 57 + 58 + 59 +))) 60 + 61 + 62 +== (% id="cke_bm_29973S" style="display:none" %) (%%)Data Analyst == 63 + 64 +>What Have You Done in This Week? 65 + 66 +((( 67 +{{success}} 68 +What Have You Done? 69 +{{/success}} 70 + 71 +1. **Dharma Dexa Phase 2** – Completed the second phase of Dharma Dexa. 72 +1. **Migration from Looker to Metabase** – Transitioning data visualization and analytics from Looker to Metabase. 73 +1. **Redesigned DKonsul Data** – Improved the structure and organization of DKonsul data. 74 +1. **Ad-hoc Requests** – Handled various on-demand data requests. 75 + 76 +{{warning}} 77 +What Issues You Have? 78 +{{/warning}} 79 + 80 +1. **Metabase Limitations** – Limited chart options and flexibility in customization, particularly a lack of aggregation functions. 81 + 82 +{{info}} 83 +What Next You Will Do? (Optional) 84 +{{/info}} 85 + 86 +1. **Continue Redesigning DKonsul Data** 87 +1*. Refining the data funnel from consultation → prescription → transaction. 88 +1. **Continue Migration to Metabase** – Ensuring a smooth transition from Looker to Metabase. 89 +1. **Dharma Dexa Phase 3** – Proceeding with the next phase of the Dharma Dexa project. 90 +1. **AppSheet MCN Visit Tracker Dashboard** – Developing and optimizing the dashboard. 91 + 92 +**What Support You Need? (Optional)** 93 + 94 +1. **Data Validation** – Ensuring data accuracy and consistency. 95 +1. **Dharma Dexa Screening Enhancements** 96 +1*. Assigning a **new screening ID** for each event, especially if different questions and inputs are involved. 97 +1*. Adding **location input (province, city)** for better analysis. 98 +))) 99 + 100 + 101 +---- 102 + 103 +== Data Analyst & AI == 104 + 105 +>What Have You Done in This Week? 106 + 107 +((( 108 +{{success}} 109 +What Have You Done? 110 +{{/success}} 111 + 112 +**AUTOMARK** 113 + 114 +1. Deploy, evaluate and making documentation of **Screening Dharma Dexa API** to be accessed by Tech team. The API is deployed at [[https:~~/~~/datalake.ptgue.com/v1/users_dd>>https://datalake.ptgue.com/v1/users_dd]] and the **documentation is already sent to Tech team**. **The API could be accessed using prompt** to retrieve user that relevant to the prompt. Per 13 Feb 2024, **the RAG Accuracy is 87,50%** 115 +1. Deploy, evaluate and making documentation of **master user GUE Ecosystem** to be accessed by Tech team. The API is deployed at [[https:~~/~~/datalake.ptgue.com/v1/users>>https://datalake.ptgue.com/v1/users_dd]] and the **documentation is already sent to Tech team**. **The API is already tested by Tech team and have no issues.** 116 + 117 +**MCN** 118 + 119 +1. Fixing resource spike issue caused by scraping schedule. The issue is already fixed and pass the test for 3 days (wednesday, thursday, and friday). **Scraping time is decreased from 1 minute per user to the maximum of +-20 seconds per user.** **(Pairing with Syifa-DE)** 120 +1. **Repoint, re-align, and redesign the pipeline and database** that being consumed for scrapping.** (Pairing with Syifa-DE)** 121 + 122 +**Automation** 123 + 124 +1. **Implement compliance mapping** with the newest data from compliance team 125 +1. Data, flow, and script validation for user report performance doctor. **The sample report is already validated by dr.Astrid** 126 + 127 +**Others** 128 + 129 +1. Reqeust + enhance dashboard merck 130 +1. Request dkonsul data for online doctors, transactions, prescriptions, Dexa prescriptions, comparison new users january 131 +1. Request data ICD-10 132 + 133 +{{warning}} 134 +What Issues You Have? 135 +{{/warning}} 136 + 137 +1. There is no info regarding database repointing of MCN from tech team. So, the data is not updated. Already solved by coordinating with Product Team 138 +1. Need to enhance the RAG accuracy to around 90% 139 + 140 +{{info}} 141 +What Next You Will Do? (Optional) 142 +{{/info}} 143 + 144 +1. Increase RAG accuracy by adding more train query LLM 145 +1. Start daily recurring scraping for tiktok dashboard 146 +1. Start dkonsul insight next week 147 + 148 +**What Support You Need? (Optional)** 149 + 150 + 151 +=== Summary === 152 + 153 +Berikut ringkasan laporan mingguan dari 14 Februari 2025 yang ditulis oleh Haekal Yusril Faizin. Laporan ini merangkum aktivitas dan isu dari tim Data Engineer, Data Analyst, dan Data Analyst & AI. 154 + 155 +**Data Analyst** 156 + 157 +* Menyelesaikan fase kedua Dharma Dexa. 158 +* Memindahkan visualisasi data dan analitik dari Looker ke Metabase. 159 +* Mendesain ulang data DKonsul untuk struktur dan organisasi yang lebih baik. 160 +* Menangani berbagai permintaan data on-demand. 161 + 162 +**Data Analyst & AI** 163 + 164 +* Menerapkan, mengevaluasi, dan mendokumentasikan API Screening Dharma Dexa untuk diakses oleh tim Tech. 165 +* Menerapkan, mengevaluasi, dan mendokumentasikan master user GUE Ecosystem untuk diakses oleh tim Tech. 166 +* Memperbaiki masalah lonjakan sumber daya yang disebabkan oleh jadwal scraping. 167 +* Merepoint, menyelaraskan ulang, dan mendesain ulang pipeline dan database yang digunakan untuk scraping. 168 +* Menerapkan pemetaan kepatuhan dengan data terbaru dari tim kepatuhan. 169 +* Memvalidasi data, alur, dan skrip untuk kinerja laporan pengguna dokter. 170 + 171 +**Isu** 172 + 173 +* Pilihan bagan dan fleksibilitas yang terbatas dalam kustomisasi Metabase. 174 +* Kurangnya informasi mengenai database repointing MCN dari tim teknologi. 175 +* Perlu meningkatkan akurasi RAG menjadi sekitar 90%. 176 + 177 +**Langkah Selanjutnya** 178 + 179 +* Melanjutkan desain ulang data DKonsul dan migrasi ke Metabase. 180 +* Melanjutkan dengan Dharma Dexa Fase 3 dan mengembangkan Dasbor Pelacak Kunjungan AppSheet MCN. 181 +* Meningkatkan akurasi RAG, memulai scraping berulang harian untuk dasbor TikTok, dan memulai wawasan DKonsul minggu depan. 182 + 183 +**Dukungan yang Dibutuhkan** 184 + 185 +* Validasi data dan peningkatan penyaringan Dharma Dexa. 186 +)))