From 41d5e1901fca36a4832d5e0adc246ba2442e8d81 Mon Sep 17 00:00:00 2001 From: shajoezhu Date: Mon, 8 Apr 2024 11:01:03 +0800 Subject: [PATCH] update --- _quarto.yml | 6 +-- docs/program.html | 38 +++++++++---------- docs/search.json | 95 ++++++++++++++++++++++++++++++++++++++++++++++- docs/sitemap.xml | 6 ++- program.qmd | 10 ++--- 5 files changed, 125 insertions(+), 30 deletions(-) diff --git a/_quarto.yml b/_quarto.yml index 42569a8..f997160 100644 --- a/_quarto.yml +++ b/_quarto.yml @@ -13,12 +13,12 @@ website: - href: index.qmd text: Home icon: "house-fill" - - href: reg.qmd - text: Registration - icon: "calendar" - href: program.qmd text: Program icon: "folder" + - href: questions.qmd + text: Q & A + icon: "question-mark" page-footer: left: | Supported by diff --git a/docs/program.html b/docs/program.html index ce02686..0d91ff1 100644 --- a/docs/program.html +++ b/docs/program.html @@ -94,14 +94,14 @@ Home @@ -163,19 +163,19 @@

Program

9:10-9:15 -Opening Remarks +Opening Remarks (video) Yan Qiao, BeiGene Beijing 9:15-9:45 - Dynamic CDISC Analysis Results Data with the R Packages {cards} and {gtsummary} + Dynamic CDISC Analysis Results Data with the R Packages {cards} and {gtsummary} (video) Daniel Sjoberg, Roche Online 9:45-10:10 -Optimize decision-making efficiency and speed by performing exploratory analysis through the MedDRAH platform +Optimize decision-making efficiency and speed by performing exploratory analysis through the MedDRAH platform (video question) Huadan Li, Zhiping Yan, Dizal Beijing @@ -187,19 +187,19 @@

Program

10:40-11:05 -Leveraging R for Real-Time Data Analysis and Reporting in the AI+HI Paradigm +Leveraging R for Real-Time Data Analysis and Reporting in the AI+HI Paradigm (video question) Hao Chen, Beigene Shanghai 11:05-11:30 -From XPT to Dataset-JSON as Alternative Transport Format for Regulatory Submission in R +From XPT to Dataset-JSON as Alternative Transport Format for Regulatory Submission in R (video) Huan Lu, Neo Yang, Sanofi Beijing 11:30-11:55 -Create TLGs and log files by sassy +Create TLGs and log files by sassy (video) Perphy Zhao, Sanofi Beijing @@ -211,25 +211,25 @@

Program

13:00-13:25 -RShiny tool: TDM automation +RShiny tool: TDM automation (video) Hao Huang, Novartis Shanghai 13:25-13:50 -Auto Subgroup TLG Generation With R Shiny +Auto Subgroup TLG Generation With R Shiny (video) Yufan Chen, J&J Shanghai 13:50-14:15 -SAS datasets comparison using R shiny app +SAS datasets comparison using R shiny app (video) Weiwei Jiao, MSD Beijing 14:15-14:40 -MediSum Shiny-App: Interactive Data Summarising using Raw Data demo +MediSum Shiny-App: Interactive Data Summarising using Raw Data demo (video) Jingya Wang, Hengrui Beijing @@ -241,31 +241,31 @@

Program

15:10-15:35 -Chevron: A open-source package to facilitate generation of outputs +Chevron: A open-source package to facilitate generation of outputs (video) Liming Li, Roche Shanghai 15:35-16:00 -R Shiny Tool for Mayo Score Monitoring in UC Study demo +R Shiny Tool for Mayo Score Monitoring in UC Study demo (video) Shaoming Liu, Pfizer Shanghai 16:00-16:25 -DaVinci Journey in the Early Phase Oncology Study +DaVinci Journey in the Early Phase Oncology Study (video 1 video 2 question) Yifan Han, Boehringer Ingelheim Shanghai 16:25-16:50 -Applying R to Big Data Processing ——Dtplyr as an Example +Applying R to Big Data Processing ——Dtplyr as an Example (video) Kaiping Yang, Beigene Online 16:50-17:00 -Closing Remarks +Closing Remarks (video) Yanli Chang, Novartis Shanghai diff --git a/docs/search.json b/docs/search.json index 07bc3b5..7db62ab 100644 --- a/docs/search.json +++ b/docs/search.json @@ -18,13 +18,104 @@ "href": "program.html", "title": "China Pharma RUG Meeting Program", "section": "", - "text": "Time\nTopic\nPresenter\nLocation\n\n\n\n\n8:30-9:00\nRegistration\n\n\n\n\n9:00-9:10\nWelcome\nJoe Zhu, Roche; Wang Zhang, MSD\nBeijing, Shanghai\n\n\n9:10-9:15\nOpening Remarks\nYan Qiao, BeiGene\nBeijing\n\n\n9:15-9:45\n Dynamic CDISC Analysis Results Data with the R Packages {cards} and {gtsummary}\nDaniel Sjoberg, Roche\nOnline\n\n\n9:45-10:10\nOptimize decision-making efficiency and speed by performing exploratory analysis through the MedDRAH platform\nHuadan Li, Zhiping Yan, Dizal\nBeijing\n\n\n10:10-10:40\nBreak\n\n\n\n\n10:40-11:05\nLeveraging R for Real-Time Data Analysis and Reporting in the AI+HI Paradigm\nHao Chen, Beigene\nShanghai\n\n\n11:05-11:30\nFrom XPT to Dataset-JSON as Alternative Transport Format for Regulatory Submission in R\nHuan Lu, Neo Yang, Sanofi\nBeijing\n\n\n11:30-11:55\nCreate TLGs and log files by sassy\nPerphy Zhao, Sanofi\nBeijing\n\n\n12:00-13:00\nLunch Break\n\n\n\n\n13:00-13:25\nRShiny tool: TDM automation\nHao Huang, Novartis\nShanghai\n\n\n13:25-13:50\nAuto Subgroup TLG Generation With R Shiny\nYufan Chen, J&J\nShanghai\n\n\n13:50-14:15\nSAS datasets comparison using R shiny app\nWeiwei Jiao, MSD\nBeijing\n\n\n14:15-14:40\nMediSum Shiny-App: Interactive Data Summarising using Raw Data demo\nJingya Wang, Hengrui\nBeijing\n\n\n14:40-15:10\nBreak\n\n\n\n\n15:10-15:35\nChevron: A open-source package to facilitate generation of outputs\nLiming Li, Roche\nShanghai\n\n\n15:35-16:00\nR Shiny Tool for Mayo Score Monitoring in UC Study demo\nShaoming Liu, Pfizer\nShanghai\n\n\n16:00-16:25\nDaVinci Journey in the Early Phase Oncology Study\nYifan Han, Boehringer Ingelheim\nShanghai\n\n\n16:25-16:50\nApplying R to Big Data Processing ——Dtplyr as an Example\nKaiping Yang, Beigene\nOnline\n\n\n16:50-17:00\nClosing Remarks\nYanli Chang, Novartis\nShanghai\n\n\n\n\n\nDaniel Sjoberg, Roche\nABSTRACT\nThe CDISC Analysis Results Data (ARD) Model is an emerging standard for encoding statistical analysis outcomes in a machine-readable format. Its primary objective is to streamline the processes of automation, ensuring reproducibility, promoting reusability, and enhancing traceability.\nThe {cards} R package offers a range of functions for ARD generation, from basic univariate summaries like means and tabulations to complex multivariable summaries encompassing regression models and statistical tests.\nThe package includes functionalities to represent results in various formats, including JSON and YAML. Thanks to its flexible structures, the {cards} package can be harnessed in diverse applications, such as generating tables for regulatory submissions and conducting quality control checks on existing tables. Furthermore, the {cards} ARD object can be accessed through a REST API, allowing writers to dynamically incorporate table results into reports.\nWhile {cards} calculates statistics and stores them in a structured object, it cannot present those results; this, however, is where the {gtsummary} package shines. The {gtsummary} package offers a modular framework to construct summary tables. It is the most widely used package for summary tables in the healthcare/pharmaceutical space, and won the American Statistical Association’s 2021 award for Innovation in Statistical Programming and Analytics. The {gtsummary} package is currently being refactored to utilize {cards} as its backend, which will allow users to both extract an ARD object from a {gtsummary} table and use an ARD object to construct a {gtsummary} table.\nThe {cards} and {gtsummary} packages stand as robust and versatile tools, poised to assist in a multitude of analytical endeavors.\n\n\n\nHuadan Li, Zhiping Yan, Dizal\nABSTRACT\nMedical Data Review and Analysis Hub (MedDRAH /’med’ra:/) is an interactive Web-based real-time clinical trial data monitoring and analysis platform that enables Medical and PV Physicians, Statisticians, and PK Scientists to review, analyze, and generate reports through intuitive point-and-click wizards. The functions below are included in MedDRAH: Review real-time EDC and SDTM data; Generate visual reports for medical data review; Perform comprehensive exploratory analysis utilizing a complete suite of analysis and reporting (A & R) tools; Create the patient narratives; Conduct standardized pharmacokinetic statistical analyses, including assessments of dose-proportionality, food/fed effects, drug-drug interactions (DDI), and exposure-response (E-R) relationships. This presentation will introduce how to design and develop such platform by using R and JavaScript.\n\n\n\nHao Chen, Beigene\nABSTRACT\nR is an excellent programming language for data analysis and reporting, featuring a mature platform developed by Posit@ and a thriving ecosystem within the pharmaceutical industry.\nBeing open-source, R offers an extensive codebase that allows AI to learn and improve. This enables the automation of R code comprehension and generation through AI, offering significant potential for the development of AI-powered R applications.\nIn parallel, R Shiny offers a superb UI framework that enhances human-computer interaction, integrating AI with human intelligence seamlessly. This presents a valuable opportunity to leverage R for real-time data analysis and reporting within the AI+HI paradigm.\nIn our presentation, we will showcase a R Shiny application designed to demonstrate this concept. We will illustrate how this integrated approach can significantly improve efficiency and decision-making in our daily work.\n\n\n\nHuan Lu, Neo Yang, Sanofi\nABSTRACT\nCDISC and PHUSE announced a new pilot project aimed at supporting the adoption of Dataset-JSON as an alternative transport format for replacing XPT as the default file format for clinical and device data submissions to regulatory authorities. New standards will be able to take advantage of enhanced capabilities, and drive efficiencies, consistency, and re-usability across the clinical research data life cycle. To leverage the features of Dataset-JSON has become the key for future development of CDISC foundational standards, this presentation will show how R can work with Dataset-JSON, both reading and writing.\n\n\n\nYifan Han, Boehringer Ingelheim\nABSTRACT\nWhat is DaVinci? DaVinci is an innovative solution at Boehringer Ingelheim to provide Data Access and dynamic Visualization for clinical insight. It enables real time access to clinical trial data and provides tools to review, aggregate and visualize data, which further fastens drug development. Apps Multiple Apps were developed and deployed under DaVinci ecosystem. Then they were put into practice and served teams for different purpose. For instance, EBAS mainly focuses on Exploratory Biomarker data Analysis, RENOVATE provides insights in terms of efficacy and Modular DaVinci paves the way for medicine team to monitor data. Moreover, DaVinci ecosystem renders flexibilities for users to include additional customized modules. RENOVATE RECIST (v1.1) based data moNitOring Visualization and Analytics Tool for Efficacy RENOVATE is an effective tool to generate evidence from POCP studies. It enables fast decision making for late phase development. It could help visualize the response data, monitor study data during the trial conduct and detect early efficacy signal, etc.\n\n\n\nHao Huang, Novartis\nABSTRACT\nThe SDTM Trial Design Model (TDM) is a required part of all CDISC SDTM electronic data submissions. Many colleagues find the TDM creation to be time-consuming and tedious due to a lot of \"copy-paste\" repetitive work. Additionally, some may feel uncomfortable with the Trial Summary (TS) domain. To enhance the experience of creating TDM for colleagues, we have developed an R ShinyApp-based automation tool, where users only need to upload the study protocol, and the tool will automatically extract the TDM information from protocol using regular expressions. Finally, the tool exports an Excel TDM report for downstream use. The tool also offers optional input fields and customized options to support the creation of a more study specific TDM report. Using this tool, users can easily obtain all the required TS parameter and most commonly used parameters for TS. It also highlights IETEST length over 200 and automatically converts special characters, etc. In practice, it takes only 10 to 40 seconds to complete the task with an acceptable accuracy rate. This tool can significantly enhance your work efficiency and improve your experience of working with TDM in daily work.\n\n\n\nYufan Chen, J&J\nABSTRACT\nWith the aim of eliminating gaps in patient access to new and innovative treatments developed in global pipeline, subgroup analysis for regulatory submission based on multi-regional clinical trial (MRCT) has been conducted more and more frequently by global pharmaceutical companies. The workflow of a subgroup analysis for a statistical programmer includes modifying existing programs from MRCT by performing analysis on a subset of the subjects’ population which involves highly repetitive manual work.   This presentation will introduce an automatic approach to modify programs in subgroup analysis. In this new workflow, statistical programmers are armed with a tool that can automate the generation of programs which are ready to be executed to create the outputs that fit the needs of regional regulators. Compared to updating programs one after another manually, this automatic tool helps to improve the efficiency and quality of subgroup analysis. Moreover, the tool can be utilized to produce subgroup versions of programs for any subgroup potentially, i.e., subgroup by sex or age groups.\n\n\n\nWeiwei Jiao, MSD\nABSTRACT\nA common task for programmers and statisticians is to compare SAS datasets in different folders, which traditionally requires the opening of SAS software and the writing of SAS code. While this is a simple process, it can be time-consuming when performed on a regular basis. The proposed solution leverages the R shiny app to compare SAS datasets, requiring only the input of the SAS dataset path and a simple click to generate comparison results, thus saving both time and effort. The paper will detail different methods of implementing this comparison via the R shiny app and will evaluate the respective advantages and disadvantages of each method.\n\n\n\nJingya Wang, Hengrui\nABSTRACT\nIn response to the increasing demand for medical reviews and the need for concise data summaries, we developed a SHINY-APP focusing on generating comprehensive tables and plots directly from EDC raw data. Currently supporting demographic tables, adverse event summaries, tumor-related efficacy analyses (e.g., KM plots, waterfall plots), and more. This tool utilizes a JSON-based spec as a guide, allowing adaptability to changes in EDC versions through spec modifications. Furthermore, our app integrates well-known R packages like rtables, tplyr, r2rtf, and Biogen’s filter module, enhancing its versatility in data exploration and analysis.\n\n\n\nLiming Li, Roche\nABSTRACT\nWithin the NEST family, we already open-sourced a lot of packages to as the backbone of clinical reporting, including rtables, tern. On top of that, we also have tlg-catalog serving as an example to facilitate the generation of outputs. However, many outputs are standard across multiple studies, and there are only minimal differences across studies. To reduce the effort to create these outputs, we developed a new package, chevron, to bring standard in a user-friendly manner. With chevron, the standards are easily implemented in a scaled manner. Chevron has already been used within Roche in multiple projects. We have also open-sourced chevron to better support you generating the outputs.\n\n\n\nShaoming Liu, Pfizer\nABSTRACT\nMayo score is the primary endpoint for UC studies. It includes sub-component of stool frequency and rectal bleeding with raw data collected in daily diary. Calculation of this sub-component for mayo score involves selection of valid days among daily diary data based on bowel preparation date and endoscopy date. It also includes sub-component of endoscopic readings from multiple readers and adjudication. UC study targeting on population with certain UC severity involves inclusion criteria pertaining to disease characteristics, such as inclusion of participants with a 5 to 9 score on the mMS (modified Mayo Score) including an endoscopy sub-score of at least 2, which requires study team to calculate mayo score at screening and/or baseline. This R shiny app is a visualized tool which facilitates study team with mayo score monitoring from various perspective and purposes, including individual daily diary quality check of valid days and sub-scores, review of endoscopic reader’s consistency, visualization of mayo score and its sub-component over time. It provides real-time monitoring to enable study team with participants inclusion verification, direct interaction with data, simplified view of complex data from multiple sources, and enhanced knowledge of patterns and trends. Excellent and plaudit feedback for this tool has been received for this tool from study team.\n\n\n\nPerphy Zhao, Sanofi\nABSTRACT\nThis paper aims to introduce a newer package called sassy that makes R easier, especially pure SAS programmers to create TLGs and logfiles. This package is a meta-package brings several SAS concept to R, and the programming grammar is highly similar with SAS, including libr which gives programmer the ability to define libnames, generate data dictionaries, and simulate data steps, fmtr which provides functions for formatting data and creating format catalogs, procs which is a set of functions that simulate SASprocedures and includes simulations of FREQ, MEANS, TRANSPOSE, SORT and PRINT procedures, reporter which is a reporting package with easy layout capabilities and the ability to write reports in RTF, DOCX, TXT and HTML file formats, logr which produces a traceable log files, common which is a set of utility functions across the sassy family packages, and often useful in their own right.\n\n\n\nKaiping Yang, Beigene\nABSTRACT\nBackground Data analysis in the pharmaceutical industry requires processing large amounts of data, such as clinical trials, genomes, drug reactions, etc. R is a widely used data science tool that provides rich data manipulation and analysis capabilities. data.table is an R package that can efficiently handle large-scale datasets, providing fast sorting, grouping, aggregation, joining, and other operations. dplyr is another R package that provides a concise data manipulation syntax, using verb functions to describe data transformations, such as filtering, selecting, arranging, grouping, summarizing, etc. dtplyr is an R package that combines data.table and dplyr, allowing you to use dplyr’s syntax to operate data.table objects, while taking advantage of data.table’s high performance.\nPurpose The purpose of this article is to introduce how to use dtplyr to process big data in the pharmaceutical industry, improving data analysis efficiency and readability. This article will show the usage of dtplyr through several examples, including data reading, data cleaning, data processing, and data analysis. This article will compare the performance and ease of use of dtplyr and other R packages, as well as the advantages, disadvantages, and application scenarios of dtplyr.\nMethods The data used in this article contains millions of rows. The R version used in this article is 4.1.2, and the R packages used are dtplyr, data.table, dplyr, tidyverse, microbenchmark, etc. The dtplyr version used in this article is 1.1.0, using the lazy_dt() function to convert data frames into lazy data table objects, then using dplyr’s verb functions to operate on the data, and finally using the as.data.table() or as_tibble() function to convert the results into data frames or tibble objects. The microbenchmark() function is used to compare the running time of dtplyr and other R packages, and the autoplot() function is used to plot the running time distribution.\nResults This article shows how to use dtplyr to complete the following tasks: Reading large files: Reading large files: using the fread() function to quickly read large files. Data cleaning: using the filter() function to filter data, using the mutate() function to create or modify variables, using the recode() function to recode variables, using the na.omit() function to delete missing values. Data processing: using the group_by() function to group data, using the summarise() function to summarize data, using the arrange() function to sort data, using the left_join() function to join data. Data analysis: using the count() function to count data, using the top_n() function to select the largest or smallest data, using the ggplot() function to plot the data distribution and relationship. This article compares the performance of dtplyr and other R packages, finding that dtplyr is faster than dplyr in most cases, but slower than data.table, especially in grouping and joining operations, data.table has a clear advantage. This article summarizes the advantages, disadvantages, and application scenarios of dtplyr, considering dtplyr as a compromise between dplyr and data.table, providing dplyr’s ease of use and data.table’s high performance, but also having some limitations, such as not supporting all dplyr functions, not supporting in-place updates, not supporting distributed computing, etc." + "text": "Time\nTopic\nPresenter\nLocation\n\n\n\n\n8:30-9:00\nRegistration\n\n\n\n\n9:00-9:10\nWelcome\nJoe Zhu, Roche; Wang Zhang, MSD\nBeijing, Shanghai\n\n\n9:10-9:15\nOpening Remarks (video)\nYan Qiao, BeiGene\nBeijing\n\n\n9:15-9:45\n Dynamic CDISC Analysis Results Data with the R Packages {cards} and {gtsummary} (video)\nDaniel Sjoberg, Roche\nOnline\n\n\n9:45-10:10\nOptimize decision-making efficiency and speed by performing exploratory analysis through the MedDRAH platform (video question)\nHuadan Li, Zhiping Yan, Dizal\nBeijing\n\n\n10:10-10:40\nBreak\n\n\n\n\n10:40-11:05\nLeveraging R for Real-Time Data Analysis and Reporting in the AI+HI Paradigm (video question)\nHao Chen, Beigene\nShanghai\n\n\n11:05-11:30\nFrom XPT to Dataset-JSON as Alternative Transport Format for Regulatory Submission in R (video)\nHuan Lu, Neo Yang, Sanofi\nBeijing\n\n\n11:30-11:55\nCreate TLGs and log files by sassy (video)\nPerphy Zhao, Sanofi\nBeijing\n\n\n12:00-13:00\nLunch Break\n\n\n\n\n13:00-13:25\nRShiny tool: TDM automation (video)\nHao Huang, Novartis\nShanghai\n\n\n13:25-13:50\nAuto Subgroup TLG Generation With R Shiny (video)\nYufan Chen, J&J\nShanghai\n\n\n13:50-14:15\nSAS datasets comparison using R shiny app (video)\nWeiwei Jiao, MSD\nBeijing\n\n\n14:15-14:40\nMediSum Shiny-App: Interactive Data Summarising using Raw Data demo (video)\nJingya Wang, Hengrui\nBeijing\n\n\n14:40-15:10\nBreak\n\n\n\n\n15:10-15:35\nChevron: A open-source package to facilitate generation of outputs (video)\nLiming Li, Roche\nShanghai\n\n\n15:35-16:00\nR Shiny Tool for Mayo Score Monitoring in UC Study demo (video)\nShaoming Liu, Pfizer\nShanghai\n\n\n16:00-16:25\nDaVinci Journey in the Early Phase Oncology Study (video 1 video 2 question)\nYifan Han, Boehringer Ingelheim\nShanghai\n\n\n16:25-16:50\nApplying R to Big Data Processing ——Dtplyr as an Example (video)\nKaiping Yang, Beigene\nOnline\n\n\n16:50-17:00\nClosing Remarks (video)\nYanli Chang, Novartis\nShanghai\n\n\n\n\n\nDaniel Sjoberg, Roche\nABSTRACT\nThe CDISC Analysis Results Data (ARD) Model is an emerging standard for encoding statistical analysis outcomes in a machine-readable format. Its primary objective is to streamline the processes of automation, ensuring reproducibility, promoting reusability, and enhancing traceability.\nThe {cards} R package offers a range of functions for ARD generation, from basic univariate summaries like means and tabulations to complex multivariable summaries encompassing regression models and statistical tests.\nThe package includes functionalities to represent results in various formats, including JSON and YAML. Thanks to its flexible structures, the {cards} package can be harnessed in diverse applications, such as generating tables for regulatory submissions and conducting quality control checks on existing tables. Furthermore, the {cards} ARD object can be accessed through a REST API, allowing writers to dynamically incorporate table results into reports.\nWhile {cards} calculates statistics and stores them in a structured object, it cannot present those results; this, however, is where the {gtsummary} package shines. The {gtsummary} package offers a modular framework to construct summary tables. It is the most widely used package for summary tables in the healthcare/pharmaceutical space, and won the American Statistical Association’s 2021 award for Innovation in Statistical Programming and Analytics. The {gtsummary} package is currently being refactored to utilize {cards} as its backend, which will allow users to both extract an ARD object from a {gtsummary} table and use an ARD object to construct a {gtsummary} table.\nThe {cards} and {gtsummary} packages stand as robust and versatile tools, poised to assist in a multitude of analytical endeavors.\n\n\n\nHuadan Li, Zhiping Yan, Dizal\nABSTRACT\nMedical Data Review and Analysis Hub (MedDRAH /’med’ra:/) is an interactive Web-based real-time clinical trial data monitoring and analysis platform that enables Medical and PV Physicians, Statisticians, and PK Scientists to review, analyze, and generate reports through intuitive point-and-click wizards. The functions below are included in MedDRAH: Review real-time EDC and SDTM data; Generate visual reports for medical data review; Perform comprehensive exploratory analysis utilizing a complete suite of analysis and reporting (A & R) tools; Create the patient narratives; Conduct standardized pharmacokinetic statistical analyses, including assessments of dose-proportionality, food/fed effects, drug-drug interactions (DDI), and exposure-response (E-R) relationships. This presentation will introduce how to design and develop such platform by using R and JavaScript.\n\n\n\nHao Chen, Beigene\nABSTRACT\nR is an excellent programming language for data analysis and reporting, featuring a mature platform developed by Posit@ and a thriving ecosystem within the pharmaceutical industry.\nBeing open-source, R offers an extensive codebase that allows AI to learn and improve. This enables the automation of R code comprehension and generation through AI, offering significant potential for the development of AI-powered R applications.\nIn parallel, R Shiny offers a superb UI framework that enhances human-computer interaction, integrating AI with human intelligence seamlessly. This presents a valuable opportunity to leverage R for real-time data analysis and reporting within the AI+HI paradigm.\nIn our presentation, we will showcase a R Shiny application designed to demonstrate this concept. We will illustrate how this integrated approach can significantly improve efficiency and decision-making in our daily work.\n\n\n\nHuan Lu, Neo Yang, Sanofi\nABSTRACT\nCDISC and PHUSE announced a new pilot project aimed at supporting the adoption of Dataset-JSON as an alternative transport format for replacing XPT as the default file format for clinical and device data submissions to regulatory authorities. New standards will be able to take advantage of enhanced capabilities, and drive efficiencies, consistency, and re-usability across the clinical research data life cycle. To leverage the features of Dataset-JSON has become the key for future development of CDISC foundational standards, this presentation will show how R can work with Dataset-JSON, both reading and writing.\n\n\n\nYifan Han, Boehringer Ingelheim\nABSTRACT\nWhat is DaVinci? DaVinci is an innovative solution at Boehringer Ingelheim to provide Data Access and dynamic Visualization for clinical insight. It enables real time access to clinical trial data and provides tools to review, aggregate and visualize data, which further fastens drug development. Apps Multiple Apps were developed and deployed under DaVinci ecosystem. Then they were put into practice and served teams for different purpose. For instance, EBAS mainly focuses on Exploratory Biomarker data Analysis, RENOVATE provides insights in terms of efficacy and Modular DaVinci paves the way for medicine team to monitor data. Moreover, DaVinci ecosystem renders flexibilities for users to include additional customized modules. RENOVATE RECIST (v1.1) based data moNitOring Visualization and Analytics Tool for Efficacy RENOVATE is an effective tool to generate evidence from POCP studies. It enables fast decision making for late phase development. It could help visualize the response data, monitor study data during the trial conduct and detect early efficacy signal, etc.\n\n\n\nHao Huang, Novartis\nABSTRACT\nThe SDTM Trial Design Model (TDM) is a required part of all CDISC SDTM electronic data submissions. Many colleagues find the TDM creation to be time-consuming and tedious due to a lot of \"copy-paste\" repetitive work. Additionally, some may feel uncomfortable with the Trial Summary (TS) domain. To enhance the experience of creating TDM for colleagues, we have developed an R ShinyApp-based automation tool, where users only need to upload the study protocol, and the tool will automatically extract the TDM information from protocol using regular expressions. Finally, the tool exports an Excel TDM report for downstream use. The tool also offers optional input fields and customized options to support the creation of a more study specific TDM report. Using this tool, users can easily obtain all the required TS parameter and most commonly used parameters for TS. It also highlights IETEST length over 200 and automatically converts special characters, etc. In practice, it takes only 10 to 40 seconds to complete the task with an acceptable accuracy rate. This tool can significantly enhance your work efficiency and improve your experience of working with TDM in daily work.\n\n\n\nYufan Chen, J&J\nABSTRACT\nWith the aim of eliminating gaps in patient access to new and innovative treatments developed in global pipeline, subgroup analysis for regulatory submission based on multi-regional clinical trial (MRCT) has been conducted more and more frequently by global pharmaceutical companies. The workflow of a subgroup analysis for a statistical programmer includes modifying existing programs from MRCT by performing analysis on a subset of the subjects’ population which involves highly repetitive manual work.   This presentation will introduce an automatic approach to modify programs in subgroup analysis. In this new workflow, statistical programmers are armed with a tool that can automate the generation of programs which are ready to be executed to create the outputs that fit the needs of regional regulators. Compared to updating programs one after another manually, this automatic tool helps to improve the efficiency and quality of subgroup analysis. Moreover, the tool can be utilized to produce subgroup versions of programs for any subgroup potentially, i.e., subgroup by sex or age groups.\n\n\n\nWeiwei Jiao, MSD\nABSTRACT\nA common task for programmers and statisticians is to compare SAS datasets in different folders, which traditionally requires the opening of SAS software and the writing of SAS code. While this is a simple process, it can be time-consuming when performed on a regular basis. The proposed solution leverages the R shiny app to compare SAS datasets, requiring only the input of the SAS dataset path and a simple click to generate comparison results, thus saving both time and effort. The paper will detail different methods of implementing this comparison via the R shiny app and will evaluate the respective advantages and disadvantages of each method.\n\n\n\nJingya Wang, Hengrui\nABSTRACT\nIn response to the increasing demand for medical reviews and the need for concise data summaries, we developed a SHINY-APP focusing on generating comprehensive tables and plots directly from EDC raw data. Currently supporting demographic tables, adverse event summaries, tumor-related efficacy analyses (e.g., KM plots, waterfall plots), and more. This tool utilizes a JSON-based spec as a guide, allowing adaptability to changes in EDC versions through spec modifications. Furthermore, our app integrates well-known R packages like rtables, tplyr, r2rtf, and Biogen’s filter module, enhancing its versatility in data exploration and analysis.\n\n\n\nLiming Li, Roche\nABSTRACT\nWithin the NEST family, we already open-sourced a lot of packages to as the backbone of clinical reporting, including rtables, tern. On top of that, we also have tlg-catalog serving as an example to facilitate the generation of outputs. However, many outputs are standard across multiple studies, and there are only minimal differences across studies. To reduce the effort to create these outputs, we developed a new package, chevron, to bring standard in a user-friendly manner. With chevron, the standards are easily implemented in a scaled manner. Chevron has already been used within Roche in multiple projects. We have also open-sourced chevron to better support you generating the outputs.\n\n\n\nShaoming Liu, Pfizer\nABSTRACT\nMayo score is the primary endpoint for UC studies. It includes sub-component of stool frequency and rectal bleeding with raw data collected in daily diary. Calculation of this sub-component for mayo score involves selection of valid days among daily diary data based on bowel preparation date and endoscopy date. It also includes sub-component of endoscopic readings from multiple readers and adjudication. UC study targeting on population with certain UC severity involves inclusion criteria pertaining to disease characteristics, such as inclusion of participants with a 5 to 9 score on the mMS (modified Mayo Score) including an endoscopy sub-score of at least 2, which requires study team to calculate mayo score at screening and/or baseline. This R shiny app is a visualized tool which facilitates study team with mayo score monitoring from various perspective and purposes, including individual daily diary quality check of valid days and sub-scores, review of endoscopic reader’s consistency, visualization of mayo score and its sub-component over time. It provides real-time monitoring to enable study team with participants inclusion verification, direct interaction with data, simplified view of complex data from multiple sources, and enhanced knowledge of patterns and trends. Excellent and plaudit feedback for this tool has been received for this tool from study team.\n\n\n\nPerphy Zhao, Sanofi\nABSTRACT\nThis paper aims to introduce a newer package called sassy that makes R easier, especially pure SAS programmers to create TLGs and logfiles. This package is a meta-package brings several SAS concept to R, and the programming grammar is highly similar with SAS, including libr which gives programmer the ability to define libnames, generate data dictionaries, and simulate data steps, fmtr which provides functions for formatting data and creating format catalogs, procs which is a set of functions that simulate SASprocedures and includes simulations of FREQ, MEANS, TRANSPOSE, SORT and PRINT procedures, reporter which is a reporting package with easy layout capabilities and the ability to write reports in RTF, DOCX, TXT and HTML file formats, logr which produces a traceable log files, common which is a set of utility functions across the sassy family packages, and often useful in their own right.\n\n\n\nKaiping Yang, Beigene\nABSTRACT\nBackground Data analysis in the pharmaceutical industry requires processing large amounts of data, such as clinical trials, genomes, drug reactions, etc. R is a widely used data science tool that provides rich data manipulation and analysis capabilities. data.table is an R package that can efficiently handle large-scale datasets, providing fast sorting, grouping, aggregation, joining, and other operations. dplyr is another R package that provides a concise data manipulation syntax, using verb functions to describe data transformations, such as filtering, selecting, arranging, grouping, summarizing, etc. dtplyr is an R package that combines data.table and dplyr, allowing you to use dplyr’s syntax to operate data.table objects, while taking advantage of data.table’s high performance.\nPurpose The purpose of this article is to introduce how to use dtplyr to process big data in the pharmaceutical industry, improving data analysis efficiency and readability. This article will show the usage of dtplyr through several examples, including data reading, data cleaning, data processing, and data analysis. This article will compare the performance and ease of use of dtplyr and other R packages, as well as the advantages, disadvantages, and application scenarios of dtplyr.\nMethods The data used in this article contains millions of rows. The R version used in this article is 4.1.2, and the R packages used are dtplyr, data.table, dplyr, tidyverse, microbenchmark, etc. The dtplyr version used in this article is 1.1.0, using the lazy_dt() function to convert data frames into lazy data table objects, then using dplyr’s verb functions to operate on the data, and finally using the as.data.table() or as_tibble() function to convert the results into data frames or tibble objects. The microbenchmark() function is used to compare the running time of dtplyr and other R packages, and the autoplot() function is used to plot the running time distribution.\nResults This article shows how to use dtplyr to complete the following tasks: Reading large files: Reading large files: using the fread() function to quickly read large files. Data cleaning: using the filter() function to filter data, using the mutate() function to create or modify variables, using the recode() function to recode variables, using the na.omit() function to delete missing values. Data processing: using the group_by() function to group data, using the summarise() function to summarize data, using the arrange() function to sort data, using the left_join() function to join data. Data analysis: using the count() function to count data, using the top_n() function to select the largest or smallest data, using the ggplot() function to plot the data distribution and relationship. This article compares the performance of dtplyr and other R packages, finding that dtplyr is faster than dplyr in most cases, but slower than data.table, especially in grouping and joining operations, data.table has a clear advantage. This article summarizes the advantages, disadvantages, and application scenarios of dtplyr, considering dtplyr as a compromise between dplyr and data.table, providing dplyr’s ease of use and data.table’s high performance, but also having some limitations, such as not supporting all dplyr functions, not supporting in-place updates, not supporting distributed computing, etc." }, { "objectID": "program.html#program", "href": "program.html#program", "title": "China Pharma RUG Meeting Program", "section": "", - "text": "Time\nTopic\nPresenter\nLocation\n\n\n\n\n8:30-9:00\nRegistration\n\n\n\n\n9:00-9:10\nWelcome\nJoe Zhu, Roche; Wang Zhang, MSD\nBeijing, Shanghai\n\n\n9:10-9:15\nOpening Remarks\nYan Qiao, BeiGene\nBeijing\n\n\n9:15-9:45\n Dynamic CDISC Analysis Results Data with the R Packages {cards} and {gtsummary}\nDaniel Sjoberg, Roche\nOnline\n\n\n9:45-10:10\nOptimize decision-making efficiency and speed by performing exploratory analysis through the MedDRAH platform\nHuadan Li, Zhiping Yan, Dizal\nBeijing\n\n\n10:10-10:40\nBreak\n\n\n\n\n10:40-11:05\nLeveraging R for Real-Time Data Analysis and Reporting in the AI+HI Paradigm\nHao Chen, Beigene\nShanghai\n\n\n11:05-11:30\nFrom XPT to Dataset-JSON as Alternative Transport Format for Regulatory Submission in R\nHuan Lu, Neo Yang, Sanofi\nBeijing\n\n\n11:30-11:55\nCreate TLGs and log files by sassy\nPerphy Zhao, Sanofi\nBeijing\n\n\n12:00-13:00\nLunch Break\n\n\n\n\n13:00-13:25\nRShiny tool: TDM automation\nHao Huang, Novartis\nShanghai\n\n\n13:25-13:50\nAuto Subgroup TLG Generation With R Shiny\nYufan Chen, J&J\nShanghai\n\n\n13:50-14:15\nSAS datasets comparison using R shiny app\nWeiwei Jiao, MSD\nBeijing\n\n\n14:15-14:40\nMediSum Shiny-App: Interactive Data Summarising using Raw Data demo\nJingya Wang, Hengrui\nBeijing\n\n\n14:40-15:10\nBreak\n\n\n\n\n15:10-15:35\nChevron: A open-source package to facilitate generation of outputs\nLiming Li, Roche\nShanghai\n\n\n15:35-16:00\nR Shiny Tool for Mayo Score Monitoring in UC Study demo\nShaoming Liu, Pfizer\nShanghai\n\n\n16:00-16:25\nDaVinci Journey in the Early Phase Oncology Study\nYifan Han, Boehringer Ingelheim\nShanghai\n\n\n16:25-16:50\nApplying R to Big Data Processing ——Dtplyr as an Example\nKaiping Yang, Beigene\nOnline\n\n\n16:50-17:00\nClosing Remarks\nYanli Chang, Novartis\nShanghai\n\n\n\n\n\nDaniel Sjoberg, Roche\nABSTRACT\nThe CDISC Analysis Results Data (ARD) Model is an emerging standard for encoding statistical analysis outcomes in a machine-readable format. Its primary objective is to streamline the processes of automation, ensuring reproducibility, promoting reusability, and enhancing traceability.\nThe {cards} R package offers a range of functions for ARD generation, from basic univariate summaries like means and tabulations to complex multivariable summaries encompassing regression models and statistical tests.\nThe package includes functionalities to represent results in various formats, including JSON and YAML. Thanks to its flexible structures, the {cards} package can be harnessed in diverse applications, such as generating tables for regulatory submissions and conducting quality control checks on existing tables. Furthermore, the {cards} ARD object can be accessed through a REST API, allowing writers to dynamically incorporate table results into reports.\nWhile {cards} calculates statistics and stores them in a structured object, it cannot present those results; this, however, is where the {gtsummary} package shines. The {gtsummary} package offers a modular framework to construct summary tables. It is the most widely used package for summary tables in the healthcare/pharmaceutical space, and won the American Statistical Association’s 2021 award for Innovation in Statistical Programming and Analytics. The {gtsummary} package is currently being refactored to utilize {cards} as its backend, which will allow users to both extract an ARD object from a {gtsummary} table and use an ARD object to construct a {gtsummary} table.\nThe {cards} and {gtsummary} packages stand as robust and versatile tools, poised to assist in a multitude of analytical endeavors.\n\n\n\nHuadan Li, Zhiping Yan, Dizal\nABSTRACT\nMedical Data Review and Analysis Hub (MedDRAH /’med’ra:/) is an interactive Web-based real-time clinical trial data monitoring and analysis platform that enables Medical and PV Physicians, Statisticians, and PK Scientists to review, analyze, and generate reports through intuitive point-and-click wizards. The functions below are included in MedDRAH: Review real-time EDC and SDTM data; Generate visual reports for medical data review; Perform comprehensive exploratory analysis utilizing a complete suite of analysis and reporting (A & R) tools; Create the patient narratives; Conduct standardized pharmacokinetic statistical analyses, including assessments of dose-proportionality, food/fed effects, drug-drug interactions (DDI), and exposure-response (E-R) relationships. This presentation will introduce how to design and develop such platform by using R and JavaScript.\n\n\n\nHao Chen, Beigene\nABSTRACT\nR is an excellent programming language for data analysis and reporting, featuring a mature platform developed by Posit@ and a thriving ecosystem within the pharmaceutical industry.\nBeing open-source, R offers an extensive codebase that allows AI to learn and improve. This enables the automation of R code comprehension and generation through AI, offering significant potential for the development of AI-powered R applications.\nIn parallel, R Shiny offers a superb UI framework that enhances human-computer interaction, integrating AI with human intelligence seamlessly. This presents a valuable opportunity to leverage R for real-time data analysis and reporting within the AI+HI paradigm.\nIn our presentation, we will showcase a R Shiny application designed to demonstrate this concept. We will illustrate how this integrated approach can significantly improve efficiency and decision-making in our daily work.\n\n\n\nHuan Lu, Neo Yang, Sanofi\nABSTRACT\nCDISC and PHUSE announced a new pilot project aimed at supporting the adoption of Dataset-JSON as an alternative transport format for replacing XPT as the default file format for clinical and device data submissions to regulatory authorities. New standards will be able to take advantage of enhanced capabilities, and drive efficiencies, consistency, and re-usability across the clinical research data life cycle. To leverage the features of Dataset-JSON has become the key for future development of CDISC foundational standards, this presentation will show how R can work with Dataset-JSON, both reading and writing.\n\n\n\nYifan Han, Boehringer Ingelheim\nABSTRACT\nWhat is DaVinci? DaVinci is an innovative solution at Boehringer Ingelheim to provide Data Access and dynamic Visualization for clinical insight. It enables real time access to clinical trial data and provides tools to review, aggregate and visualize data, which further fastens drug development. Apps Multiple Apps were developed and deployed under DaVinci ecosystem. Then they were put into practice and served teams for different purpose. For instance, EBAS mainly focuses on Exploratory Biomarker data Analysis, RENOVATE provides insights in terms of efficacy and Modular DaVinci paves the way for medicine team to monitor data. Moreover, DaVinci ecosystem renders flexibilities for users to include additional customized modules. RENOVATE RECIST (v1.1) based data moNitOring Visualization and Analytics Tool for Efficacy RENOVATE is an effective tool to generate evidence from POCP studies. It enables fast decision making for late phase development. It could help visualize the response data, monitor study data during the trial conduct and detect early efficacy signal, etc.\n\n\n\nHao Huang, Novartis\nABSTRACT\nThe SDTM Trial Design Model (TDM) is a required part of all CDISC SDTM electronic data submissions. Many colleagues find the TDM creation to be time-consuming and tedious due to a lot of \"copy-paste\" repetitive work. Additionally, some may feel uncomfortable with the Trial Summary (TS) domain. To enhance the experience of creating TDM for colleagues, we have developed an R ShinyApp-based automation tool, where users only need to upload the study protocol, and the tool will automatically extract the TDM information from protocol using regular expressions. Finally, the tool exports an Excel TDM report for downstream use. The tool also offers optional input fields and customized options to support the creation of a more study specific TDM report. Using this tool, users can easily obtain all the required TS parameter and most commonly used parameters for TS. It also highlights IETEST length over 200 and automatically converts special characters, etc. In practice, it takes only 10 to 40 seconds to complete the task with an acceptable accuracy rate. This tool can significantly enhance your work efficiency and improve your experience of working with TDM in daily work.\n\n\n\nYufan Chen, J&J\nABSTRACT\nWith the aim of eliminating gaps in patient access to new and innovative treatments developed in global pipeline, subgroup analysis for regulatory submission based on multi-regional clinical trial (MRCT) has been conducted more and more frequently by global pharmaceutical companies. The workflow of a subgroup analysis for a statistical programmer includes modifying existing programs from MRCT by performing analysis on a subset of the subjects’ population which involves highly repetitive manual work.   This presentation will introduce an automatic approach to modify programs in subgroup analysis. In this new workflow, statistical programmers are armed with a tool that can automate the generation of programs which are ready to be executed to create the outputs that fit the needs of regional regulators. Compared to updating programs one after another manually, this automatic tool helps to improve the efficiency and quality of subgroup analysis. Moreover, the tool can be utilized to produce subgroup versions of programs for any subgroup potentially, i.e., subgroup by sex or age groups.\n\n\n\nWeiwei Jiao, MSD\nABSTRACT\nA common task for programmers and statisticians is to compare SAS datasets in different folders, which traditionally requires the opening of SAS software and the writing of SAS code. While this is a simple process, it can be time-consuming when performed on a regular basis. The proposed solution leverages the R shiny app to compare SAS datasets, requiring only the input of the SAS dataset path and a simple click to generate comparison results, thus saving both time and effort. The paper will detail different methods of implementing this comparison via the R shiny app and will evaluate the respective advantages and disadvantages of each method.\n\n\n\nJingya Wang, Hengrui\nABSTRACT\nIn response to the increasing demand for medical reviews and the need for concise data summaries, we developed a SHINY-APP focusing on generating comprehensive tables and plots directly from EDC raw data. Currently supporting demographic tables, adverse event summaries, tumor-related efficacy analyses (e.g., KM plots, waterfall plots), and more. This tool utilizes a JSON-based spec as a guide, allowing adaptability to changes in EDC versions through spec modifications. Furthermore, our app integrates well-known R packages like rtables, tplyr, r2rtf, and Biogen’s filter module, enhancing its versatility in data exploration and analysis.\n\n\n\nLiming Li, Roche\nABSTRACT\nWithin the NEST family, we already open-sourced a lot of packages to as the backbone of clinical reporting, including rtables, tern. On top of that, we also have tlg-catalog serving as an example to facilitate the generation of outputs. However, many outputs are standard across multiple studies, and there are only minimal differences across studies. To reduce the effort to create these outputs, we developed a new package, chevron, to bring standard in a user-friendly manner. With chevron, the standards are easily implemented in a scaled manner. Chevron has already been used within Roche in multiple projects. We have also open-sourced chevron to better support you generating the outputs.\n\n\n\nShaoming Liu, Pfizer\nABSTRACT\nMayo score is the primary endpoint for UC studies. It includes sub-component of stool frequency and rectal bleeding with raw data collected in daily diary. Calculation of this sub-component for mayo score involves selection of valid days among daily diary data based on bowel preparation date and endoscopy date. It also includes sub-component of endoscopic readings from multiple readers and adjudication. UC study targeting on population with certain UC severity involves inclusion criteria pertaining to disease characteristics, such as inclusion of participants with a 5 to 9 score on the mMS (modified Mayo Score) including an endoscopy sub-score of at least 2, which requires study team to calculate mayo score at screening and/or baseline. This R shiny app is a visualized tool which facilitates study team with mayo score monitoring from various perspective and purposes, including individual daily diary quality check of valid days and sub-scores, review of endoscopic reader’s consistency, visualization of mayo score and its sub-component over time. It provides real-time monitoring to enable study team with participants inclusion verification, direct interaction with data, simplified view of complex data from multiple sources, and enhanced knowledge of patterns and trends. Excellent and plaudit feedback for this tool has been received for this tool from study team.\n\n\n\nPerphy Zhao, Sanofi\nABSTRACT\nThis paper aims to introduce a newer package called sassy that makes R easier, especially pure SAS programmers to create TLGs and logfiles. This package is a meta-package brings several SAS concept to R, and the programming grammar is highly similar with SAS, including libr which gives programmer the ability to define libnames, generate data dictionaries, and simulate data steps, fmtr which provides functions for formatting data and creating format catalogs, procs which is a set of functions that simulate SASprocedures and includes simulations of FREQ, MEANS, TRANSPOSE, SORT and PRINT procedures, reporter which is a reporting package with easy layout capabilities and the ability to write reports in RTF, DOCX, TXT and HTML file formats, logr which produces a traceable log files, common which is a set of utility functions across the sassy family packages, and often useful in their own right.\n\n\n\nKaiping Yang, Beigene\nABSTRACT\nBackground Data analysis in the pharmaceutical industry requires processing large amounts of data, such as clinical trials, genomes, drug reactions, etc. R is a widely used data science tool that provides rich data manipulation and analysis capabilities. data.table is an R package that can efficiently handle large-scale datasets, providing fast sorting, grouping, aggregation, joining, and other operations. dplyr is another R package that provides a concise data manipulation syntax, using verb functions to describe data transformations, such as filtering, selecting, arranging, grouping, summarizing, etc. dtplyr is an R package that combines data.table and dplyr, allowing you to use dplyr’s syntax to operate data.table objects, while taking advantage of data.table’s high performance.\nPurpose The purpose of this article is to introduce how to use dtplyr to process big data in the pharmaceutical industry, improving data analysis efficiency and readability. This article will show the usage of dtplyr through several examples, including data reading, data cleaning, data processing, and data analysis. This article will compare the performance and ease of use of dtplyr and other R packages, as well as the advantages, disadvantages, and application scenarios of dtplyr.\nMethods The data used in this article contains millions of rows. The R version used in this article is 4.1.2, and the R packages used are dtplyr, data.table, dplyr, tidyverse, microbenchmark, etc. The dtplyr version used in this article is 1.1.0, using the lazy_dt() function to convert data frames into lazy data table objects, then using dplyr’s verb functions to operate on the data, and finally using the as.data.table() or as_tibble() function to convert the results into data frames or tibble objects. The microbenchmark() function is used to compare the running time of dtplyr and other R packages, and the autoplot() function is used to plot the running time distribution.\nResults This article shows how to use dtplyr to complete the following tasks: Reading large files: Reading large files: using the fread() function to quickly read large files. Data cleaning: using the filter() function to filter data, using the mutate() function to create or modify variables, using the recode() function to recode variables, using the na.omit() function to delete missing values. Data processing: using the group_by() function to group data, using the summarise() function to summarize data, using the arrange() function to sort data, using the left_join() function to join data. Data analysis: using the count() function to count data, using the top_n() function to select the largest or smallest data, using the ggplot() function to plot the data distribution and relationship. This article compares the performance of dtplyr and other R packages, finding that dtplyr is faster than dplyr in most cases, but slower than data.table, especially in grouping and joining operations, data.table has a clear advantage. This article summarizes the advantages, disadvantages, and application scenarios of dtplyr, considering dtplyr as a compromise between dplyr and data.table, providing dplyr’s ease of use and data.table’s high performance, but also having some limitations, such as not supporting all dplyr functions, not supporting in-place updates, not supporting distributed computing, etc." + "text": "Time\nTopic\nPresenter\nLocation\n\n\n\n\n8:30-9:00\nRegistration\n\n\n\n\n9:00-9:10\nWelcome\nJoe Zhu, Roche; Wang Zhang, MSD\nBeijing, Shanghai\n\n\n9:10-9:15\nOpening Remarks (video)\nYan Qiao, BeiGene\nBeijing\n\n\n9:15-9:45\n Dynamic CDISC Analysis Results Data with the R Packages {cards} and {gtsummary} (video)\nDaniel Sjoberg, Roche\nOnline\n\n\n9:45-10:10\nOptimize decision-making efficiency and speed by performing exploratory analysis through the MedDRAH platform (video question)\nHuadan Li, Zhiping Yan, Dizal\nBeijing\n\n\n10:10-10:40\nBreak\n\n\n\n\n10:40-11:05\nLeveraging R for Real-Time Data Analysis and Reporting in the AI+HI Paradigm (video question)\nHao Chen, Beigene\nShanghai\n\n\n11:05-11:30\nFrom XPT to Dataset-JSON as Alternative Transport Format for Regulatory Submission in R (video)\nHuan Lu, Neo Yang, Sanofi\nBeijing\n\n\n11:30-11:55\nCreate TLGs and log files by sassy (video)\nPerphy Zhao, Sanofi\nBeijing\n\n\n12:00-13:00\nLunch Break\n\n\n\n\n13:00-13:25\nRShiny tool: TDM automation (video)\nHao Huang, Novartis\nShanghai\n\n\n13:25-13:50\nAuto Subgroup TLG Generation With R Shiny (video)\nYufan Chen, J&J\nShanghai\n\n\n13:50-14:15\nSAS datasets comparison using R shiny app (video)\nWeiwei Jiao, MSD\nBeijing\n\n\n14:15-14:40\nMediSum Shiny-App: Interactive Data Summarising using Raw Data demo (video)\nJingya Wang, Hengrui\nBeijing\n\n\n14:40-15:10\nBreak\n\n\n\n\n15:10-15:35\nChevron: A open-source package to facilitate generation of outputs (video)\nLiming Li, Roche\nShanghai\n\n\n15:35-16:00\nR Shiny Tool for Mayo Score Monitoring in UC Study demo (video)\nShaoming Liu, Pfizer\nShanghai\n\n\n16:00-16:25\nDaVinci Journey in the Early Phase Oncology Study (video 1 video 2 question)\nYifan Han, Boehringer Ingelheim\nShanghai\n\n\n16:25-16:50\nApplying R to Big Data Processing ——Dtplyr as an Example (video)\nKaiping Yang, Beigene\nOnline\n\n\n16:50-17:00\nClosing Remarks (video)\nYanli Chang, Novartis\nShanghai\n\n\n\n\n\nDaniel Sjoberg, Roche\nABSTRACT\nThe CDISC Analysis Results Data (ARD) Model is an emerging standard for encoding statistical analysis outcomes in a machine-readable format. Its primary objective is to streamline the processes of automation, ensuring reproducibility, promoting reusability, and enhancing traceability.\nThe {cards} R package offers a range of functions for ARD generation, from basic univariate summaries like means and tabulations to complex multivariable summaries encompassing regression models and statistical tests.\nThe package includes functionalities to represent results in various formats, including JSON and YAML. Thanks to its flexible structures, the {cards} package can be harnessed in diverse applications, such as generating tables for regulatory submissions and conducting quality control checks on existing tables. Furthermore, the {cards} ARD object can be accessed through a REST API, allowing writers to dynamically incorporate table results into reports.\nWhile {cards} calculates statistics and stores them in a structured object, it cannot present those results; this, however, is where the {gtsummary} package shines. The {gtsummary} package offers a modular framework to construct summary tables. It is the most widely used package for summary tables in the healthcare/pharmaceutical space, and won the American Statistical Association’s 2021 award for Innovation in Statistical Programming and Analytics. The {gtsummary} package is currently being refactored to utilize {cards} as its backend, which will allow users to both extract an ARD object from a {gtsummary} table and use an ARD object to construct a {gtsummary} table.\nThe {cards} and {gtsummary} packages stand as robust and versatile tools, poised to assist in a multitude of analytical endeavors.\n\n\n\nHuadan Li, Zhiping Yan, Dizal\nABSTRACT\nMedical Data Review and Analysis Hub (MedDRAH /’med’ra:/) is an interactive Web-based real-time clinical trial data monitoring and analysis platform that enables Medical and PV Physicians, Statisticians, and PK Scientists to review, analyze, and generate reports through intuitive point-and-click wizards. The functions below are included in MedDRAH: Review real-time EDC and SDTM data; Generate visual reports for medical data review; Perform comprehensive exploratory analysis utilizing a complete suite of analysis and reporting (A & R) tools; Create the patient narratives; Conduct standardized pharmacokinetic statistical analyses, including assessments of dose-proportionality, food/fed effects, drug-drug interactions (DDI), and exposure-response (E-R) relationships. This presentation will introduce how to design and develop such platform by using R and JavaScript.\n\n\n\nHao Chen, Beigene\nABSTRACT\nR is an excellent programming language for data analysis and reporting, featuring a mature platform developed by Posit@ and a thriving ecosystem within the pharmaceutical industry.\nBeing open-source, R offers an extensive codebase that allows AI to learn and improve. This enables the automation of R code comprehension and generation through AI, offering significant potential for the development of AI-powered R applications.\nIn parallel, R Shiny offers a superb UI framework that enhances human-computer interaction, integrating AI with human intelligence seamlessly. This presents a valuable opportunity to leverage R for real-time data analysis and reporting within the AI+HI paradigm.\nIn our presentation, we will showcase a R Shiny application designed to demonstrate this concept. We will illustrate how this integrated approach can significantly improve efficiency and decision-making in our daily work.\n\n\n\nHuan Lu, Neo Yang, Sanofi\nABSTRACT\nCDISC and PHUSE announced a new pilot project aimed at supporting the adoption of Dataset-JSON as an alternative transport format for replacing XPT as the default file format for clinical and device data submissions to regulatory authorities. New standards will be able to take advantage of enhanced capabilities, and drive efficiencies, consistency, and re-usability across the clinical research data life cycle. To leverage the features of Dataset-JSON has become the key for future development of CDISC foundational standards, this presentation will show how R can work with Dataset-JSON, both reading and writing.\n\n\n\nYifan Han, Boehringer Ingelheim\nABSTRACT\nWhat is DaVinci? DaVinci is an innovative solution at Boehringer Ingelheim to provide Data Access and dynamic Visualization for clinical insight. It enables real time access to clinical trial data and provides tools to review, aggregate and visualize data, which further fastens drug development. Apps Multiple Apps were developed and deployed under DaVinci ecosystem. Then they were put into practice and served teams for different purpose. For instance, EBAS mainly focuses on Exploratory Biomarker data Analysis, RENOVATE provides insights in terms of efficacy and Modular DaVinci paves the way for medicine team to monitor data. Moreover, DaVinci ecosystem renders flexibilities for users to include additional customized modules. RENOVATE RECIST (v1.1) based data moNitOring Visualization and Analytics Tool for Efficacy RENOVATE is an effective tool to generate evidence from POCP studies. It enables fast decision making for late phase development. It could help visualize the response data, monitor study data during the trial conduct and detect early efficacy signal, etc.\n\n\n\nHao Huang, Novartis\nABSTRACT\nThe SDTM Trial Design Model (TDM) is a required part of all CDISC SDTM electronic data submissions. Many colleagues find the TDM creation to be time-consuming and tedious due to a lot of \"copy-paste\" repetitive work. Additionally, some may feel uncomfortable with the Trial Summary (TS) domain. To enhance the experience of creating TDM for colleagues, we have developed an R ShinyApp-based automation tool, where users only need to upload the study protocol, and the tool will automatically extract the TDM information from protocol using regular expressions. Finally, the tool exports an Excel TDM report for downstream use. The tool also offers optional input fields and customized options to support the creation of a more study specific TDM report. Using this tool, users can easily obtain all the required TS parameter and most commonly used parameters for TS. It also highlights IETEST length over 200 and automatically converts special characters, etc. In practice, it takes only 10 to 40 seconds to complete the task with an acceptable accuracy rate. This tool can significantly enhance your work efficiency and improve your experience of working with TDM in daily work.\n\n\n\nYufan Chen, J&J\nABSTRACT\nWith the aim of eliminating gaps in patient access to new and innovative treatments developed in global pipeline, subgroup analysis for regulatory submission based on multi-regional clinical trial (MRCT) has been conducted more and more frequently by global pharmaceutical companies. The workflow of a subgroup analysis for a statistical programmer includes modifying existing programs from MRCT by performing analysis on a subset of the subjects’ population which involves highly repetitive manual work.   This presentation will introduce an automatic approach to modify programs in subgroup analysis. In this new workflow, statistical programmers are armed with a tool that can automate the generation of programs which are ready to be executed to create the outputs that fit the needs of regional regulators. Compared to updating programs one after another manually, this automatic tool helps to improve the efficiency and quality of subgroup analysis. Moreover, the tool can be utilized to produce subgroup versions of programs for any subgroup potentially, i.e., subgroup by sex or age groups.\n\n\n\nWeiwei Jiao, MSD\nABSTRACT\nA common task for programmers and statisticians is to compare SAS datasets in different folders, which traditionally requires the opening of SAS software and the writing of SAS code. While this is a simple process, it can be time-consuming when performed on a regular basis. The proposed solution leverages the R shiny app to compare SAS datasets, requiring only the input of the SAS dataset path and a simple click to generate comparison results, thus saving both time and effort. The paper will detail different methods of implementing this comparison via the R shiny app and will evaluate the respective advantages and disadvantages of each method.\n\n\n\nJingya Wang, Hengrui\nABSTRACT\nIn response to the increasing demand for medical reviews and the need for concise data summaries, we developed a SHINY-APP focusing on generating comprehensive tables and plots directly from EDC raw data. Currently supporting demographic tables, adverse event summaries, tumor-related efficacy analyses (e.g., KM plots, waterfall plots), and more. This tool utilizes a JSON-based spec as a guide, allowing adaptability to changes in EDC versions through spec modifications. Furthermore, our app integrates well-known R packages like rtables, tplyr, r2rtf, and Biogen’s filter module, enhancing its versatility in data exploration and analysis.\n\n\n\nLiming Li, Roche\nABSTRACT\nWithin the NEST family, we already open-sourced a lot of packages to as the backbone of clinical reporting, including rtables, tern. On top of that, we also have tlg-catalog serving as an example to facilitate the generation of outputs. However, many outputs are standard across multiple studies, and there are only minimal differences across studies. To reduce the effort to create these outputs, we developed a new package, chevron, to bring standard in a user-friendly manner. With chevron, the standards are easily implemented in a scaled manner. Chevron has already been used within Roche in multiple projects. We have also open-sourced chevron to better support you generating the outputs.\n\n\n\nShaoming Liu, Pfizer\nABSTRACT\nMayo score is the primary endpoint for UC studies. It includes sub-component of stool frequency and rectal bleeding with raw data collected in daily diary. Calculation of this sub-component for mayo score involves selection of valid days among daily diary data based on bowel preparation date and endoscopy date. It also includes sub-component of endoscopic readings from multiple readers and adjudication. UC study targeting on population with certain UC severity involves inclusion criteria pertaining to disease characteristics, such as inclusion of participants with a 5 to 9 score on the mMS (modified Mayo Score) including an endoscopy sub-score of at least 2, which requires study team to calculate mayo score at screening and/or baseline. This R shiny app is a visualized tool which facilitates study team with mayo score monitoring from various perspective and purposes, including individual daily diary quality check of valid days and sub-scores, review of endoscopic reader’s consistency, visualization of mayo score and its sub-component over time. It provides real-time monitoring to enable study team with participants inclusion verification, direct interaction with data, simplified view of complex data from multiple sources, and enhanced knowledge of patterns and trends. Excellent and plaudit feedback for this tool has been received for this tool from study team.\n\n\n\nPerphy Zhao, Sanofi\nABSTRACT\nThis paper aims to introduce a newer package called sassy that makes R easier, especially pure SAS programmers to create TLGs and logfiles. This package is a meta-package brings several SAS concept to R, and the programming grammar is highly similar with SAS, including libr which gives programmer the ability to define libnames, generate data dictionaries, and simulate data steps, fmtr which provides functions for formatting data and creating format catalogs, procs which is a set of functions that simulate SASprocedures and includes simulations of FREQ, MEANS, TRANSPOSE, SORT and PRINT procedures, reporter which is a reporting package with easy layout capabilities and the ability to write reports in RTF, DOCX, TXT and HTML file formats, logr which produces a traceable log files, common which is a set of utility functions across the sassy family packages, and often useful in their own right.\n\n\n\nKaiping Yang, Beigene\nABSTRACT\nBackground Data analysis in the pharmaceutical industry requires processing large amounts of data, such as clinical trials, genomes, drug reactions, etc. R is a widely used data science tool that provides rich data manipulation and analysis capabilities. data.table is an R package that can efficiently handle large-scale datasets, providing fast sorting, grouping, aggregation, joining, and other operations. dplyr is another R package that provides a concise data manipulation syntax, using verb functions to describe data transformations, such as filtering, selecting, arranging, grouping, summarizing, etc. dtplyr is an R package that combines data.table and dplyr, allowing you to use dplyr’s syntax to operate data.table objects, while taking advantage of data.table’s high performance.\nPurpose The purpose of this article is to introduce how to use dtplyr to process big data in the pharmaceutical industry, improving data analysis efficiency and readability. This article will show the usage of dtplyr through several examples, including data reading, data cleaning, data processing, and data analysis. This article will compare the performance and ease of use of dtplyr and other R packages, as well as the advantages, disadvantages, and application scenarios of dtplyr.\nMethods The data used in this article contains millions of rows. The R version used in this article is 4.1.2, and the R packages used are dtplyr, data.table, dplyr, tidyverse, microbenchmark, etc. The dtplyr version used in this article is 1.1.0, using the lazy_dt() function to convert data frames into lazy data table objects, then using dplyr’s verb functions to operate on the data, and finally using the as.data.table() or as_tibble() function to convert the results into data frames or tibble objects. The microbenchmark() function is used to compare the running time of dtplyr and other R packages, and the autoplot() function is used to plot the running time distribution.\nResults This article shows how to use dtplyr to complete the following tasks: Reading large files: Reading large files: using the fread() function to quickly read large files. Data cleaning: using the filter() function to filter data, using the mutate() function to create or modify variables, using the recode() function to recode variables, using the na.omit() function to delete missing values. Data processing: using the group_by() function to group data, using the summarise() function to summarize data, using the arrange() function to sort data, using the left_join() function to join data. Data analysis: using the count() function to count data, using the top_n() function to select the largest or smallest data, using the ggplot() function to plot the data distribution and relationship. This article compares the performance of dtplyr and other R packages, finding that dtplyr is faster than dplyr in most cases, but slower than data.table, especially in grouping and joining operations, data.table has a clear advantage. This article summarizes the advantages, disadvantages, and application scenarios of dtplyr, considering dtplyr as a compromise between dplyr and data.table, providing dplyr’s ease of use and data.table’s high performance, but also having some limitations, such as not supporting all dplyr functions, not supporting in-place updates, not supporting distributed computing, etc." + }, + { + "objectID": "questions.html", + "href": "questions.html", + "title": "Follow-up Q & A", + "section": "", + "text": "Hao:适用人群比较广泛,有数据分析,做table,画图等需求的部门都可以,如DM 用来做data query, Programming和Stat用来计算和画图\n\n\n\nHao:当前主要是copilot,人机结合,人工验证。 主要用GPT3.5,简单到中等任务可以,复杂的需要人在draft的code上做进一步修改。GPT4 准确率会更高,但是太慢不能real-time,人机结合体验不好,同时token cost也比较高。随着LLM不断推陈出新,可以期待越来越多的任务会变得容易\n\n\n\nHao:现在是基于直接prompt和RAG辅助。之前有用到复杂的COT和GPT4去处理复杂的问题,但是处理时间太长。为了real-time达到copilot的效果,主要是用GPT3.5; 未来会考虑fine-tune,效率可以进一步提高" + }, + { + "objectID": "questions.html#kai-gu-请问下这个tool的适用人群包括哪些部门", + "href": "questions.html#kai-gu-请问下这个tool的适用人群包括哪些部门", + "title": "Follow-up Q & A", + "section": "", + "text": "Hao:适用人群比较广泛,有数据分析,做table,画图等需求的部门都可以,如DM 用来做data query, Programming和Stat用来计算和画图" + }, + { + "objectID": "questions.html#longfei-正如hao所说-我们这一行是很严格的-这个case-准确率如何", + "href": "questions.html#longfei-正如hao所说-我们这一行是很严格的-这个case-准确率如何", + "title": "Follow-up Q & A", + "section": "", + "text": "Hao:当前主要是copilot,人机结合,人工验证。 主要用GPT3.5,简单到中等任务可以,复杂的需要人在draft的code上做进一步修改。GPT4 准确率会更高,但是太慢不能real-time,人机结合体验不好,同时token cost也比较高。随着LLM不断推陈出新,可以期待越来越多的任务会变得容易" + }, + { + "objectID": "questions.html#zou-ivan-请问以上结果都是在模型中基于prompt-input做出来的吗-是否有用到其他rag或者fine-tuned", + "href": "questions.html#zou-ivan-请问以上结果都是在模型中基于prompt-input做出来的吗-是否有用到其他rag或者fine-tuned", + "title": "Follow-up Q & A", + "section": "", + "text": "Hao:现在是基于直接prompt和RAG辅助。之前有用到复杂的COT和GPT4去处理复杂的问题,但是处理时间太长。为了real-time达到copilot的效果,主要是用GPT3.5; 未来会考虑fine-tune,效率可以进一步提高" + }, + { + "objectID": "questions.html#请问和teal的融合时需要把data都按照teal做预处理吗teal_data", + "href": "questions.html#请问和teal的融合时需要把data都按照teal做预处理吗teal_data", + "title": "Follow-up Q & A", + "section": "请问和teal的融合时,需要把data都按照teal做预处理吗?(teal_data?)", + "text": "请问和teal的融合时,需要把data都按照teal做预处理吗?(teal_data?)\n是的,需要的。Teal是基于ADaM,遵守Roche内部的ADS 标准。当我们的ADaM和Roche ADaM 有出入的时候,会有报错。\n有时候提示信息也不会很明显,这样导致了的很多时间都花在了debug改数据上。" + }, + { + "objectID": "questions.html#qike-请问这个网站是开放还是仅供内部使用的", + "href": "questions.html#qike-请问这个网站是开放还是仅供内部使用的", + "title": "Follow-up Q & A", + "section": "Qike: 请问这个网站是开放还是仅供内部使用的?", + "text": "Qike: 请问这个网站是开放还是仅供内部使用的?\nZY: 目前仅仅是内部使用" + }, + { + "objectID": "questions.html#wenlu-请问ongoing-的数据如果遇到data-issue是否会影响展示的数据", + "href": "questions.html#wenlu-请问ongoing-的数据如果遇到data-issue是否会影响展示的数据", + "title": "Follow-up Q & A", + "section": "wenlu: 请问ongoing 的数据如果遇到data issue,是否会影响展示的数据。", + "text": "wenlu: 请问ongoing 的数据如果遇到data issue,是否会影响展示的数据。\nZY: 要看是什么样的data issue,是否影响了后端的计算。如果程序做了error处理,也会报出来一些错误。通过错误看是改程序还是改数据。对于不影响展示的数据,比如outlier之类的,这个工具也可以用来发现data issue" + }, + { + "objectID": "questions.html#wenlu-请问开发这个交互平台用了多长的时间和投入了多少resource", + "href": "questions.html#wenlu-请问开发这个交互平台用了多长的时间和投入了多少resource", + "title": "Follow-up Q & A", + "section": "wenlu: 请问开发这个交互平台用了多长的时间和投入了多少resource", + "text": "wenlu: 请问开发这个交互平台用了多长的时间和投入了多少resource\n不算前期调研,MDR 板块用了大概半年时间。我做了大部分的后端和前端,一名IT同事做了一部分前端,两个本组的同事帮忙做了两个后端API 的函数。" + }, + { + "objectID": "questions.html#keyu-chen-请问这个平台的web应用是支持多线程的吗会存在高并发的问题吗", + "href": "questions.html#keyu-chen-请问这个平台的web应用是支持多线程的吗会存在高并发的问题吗", + "title": "Follow-up Q & A", + "section": "Keyu Chen: 请问这个平台的Web应用是支持多线程的吗?会存在高并发的问题吗", + "text": "Keyu Chen: 请问这个平台的Web应用是支持多线程的吗?会存在高并发的问题吗\nZY: 支持多线程, 高并发还要后面去解决。现在只是先把功能做好,能用起来。" + }, + { + "objectID": "questions.html#zou-ivan-请问在build前端界面时由于后端是使用的r和python那么前端不直接用python-flask或者r-shiny构建的理由-同样对于table和plot的display有考虑直接使用python或者r中的库去实现相同效果吗", + "href": "questions.html#zou-ivan-请问在build前端界面时由于后端是使用的r和python那么前端不直接用python-flask或者r-shiny构建的理由-同样对于table和plot的display有考虑直接使用python或者r中的库去实现相同效果吗", + "title": "Follow-up Q & A", + "section": "Zou, Ivan: 请问在build前端界面时,由于后端是使用的R和python,那么前端不直接用Python Flask或者R shiny构建的理由? 同样对于table和plot的display,有考虑直接使用Python或者R中的库去实现相同效果吗?", + "text": "Zou, Ivan: 请问在build前端界面时,由于后端是使用的R和python,那么前端不直接用Python Flask或者R shiny构建的理由? 同样对于table和plot的display,有考虑直接使用Python或者R中的库去实现相同效果吗?\nZY: Javascript 做前端会更灵活和方便。互动性也会更好。Plotly 也有Python 版本吧。R 产生的图不支持互动。" + }, + { + "objectID": "questions.html#leveraging-r-for-real-time-data-analysis-and-reporting-in-the-aihi-paradigm", + "href": "questions.html#leveraging-r-for-real-time-data-analysis-and-reporting-in-the-aihi-paradigm", + "title": "Follow-up Q & A", + "section": "", + "text": "Hao:适用人群比较广泛,有数据分析,做table,画图等需求的部门都可以,如DM 用来做data query, Programming和Stat用来计算和画图\n\n\n\nHao:当前主要是copilot,人机结合,人工验证。 主要用GPT3.5,简单到中等任务可以,复杂的需要人在draft的code上做进一步修改。GPT4 准确率会更高,但是太慢不能real-time,人机结合体验不好,同时token cost也比较高。随着LLM不断推陈出新,可以期待越来越多的任务会变得容易\n\n\n\nHao:现在是基于直接prompt和RAG辅助。之前有用到复杂的COT和GPT4去处理复杂的问题,但是处理时间太长。为了real-time达到copilot的效果,主要是用GPT3.5; 未来会考虑fine-tune,效率可以进一步提高" + }, + { + "objectID": "questions.html#davinci-journey-in-the-early-phase-oncology-study", + "href": "questions.html#davinci-journey-in-the-early-phase-oncology-study", + "title": "Follow-up Q & A", + "section": "DaVinci Journey in the Early Phase Oncology Study", + "text": "DaVinci Journey in the Early Phase Oncology Study\n\n请问和teal的融合时,需要把data都按照teal做预处理吗?(teal_data?)\n是的,需要的。Teal是基于ADaM,遵守Roche内部的ADS 标准。当我们的ADaM和Roche ADaM 有出入的时候,会有报错。\n有时候提示信息也不会很明显,这样导致了的很多时间都花在了debug改数据上。" + }, + { + "objectID": "questions.html#optimize-decision-making-efficiency-and-speed-by-performing-exploratory-analysis-through-the-meddrah-platform", + "href": "questions.html#optimize-decision-making-efficiency-and-speed-by-performing-exploratory-analysis-through-the-meddrah-platform", + "title": "Follow-up Q & A", + "section": "Optimize decision-making efficiency and speed by performing exploratory analysis through the MedDRAH platform", + "text": "Optimize decision-making efficiency and speed by performing exploratory analysis through the MedDRAH platform\n\nQike: 请问这个网站是开放还是仅供内部使用的?\nZY: 目前仅仅是内部使用\n\n\nwenlu: 请问ongoing 的数据如果遇到data issue,是否会影响展示的数据。\nZY: 要看是什么样的data issue,是否影响了后端的计算。如果程序做了error处理,也会报出来一些错误。通过错误看是改程序还是改数据。对于不影响展示的数据,比如outlier之类的,这个工具也可以用来发现data issue\n\n\nwenlu: 请问开发这个交互平台用了多长的时间和投入了多少resource\n不算前期调研,MDR 板块用了大概半年时间。我做了大部分的后端和前端,一名IT同事做了一部分前端,两个本组的同事帮忙做了两个后端API 的函数。\n\n\nKeyu Chen: 请问这个平台的Web应用是支持多线程的吗?会存在高并发的问题吗\nZY: 支持多线程, 高并发还要后面去解决。现在只是先把功能做好,能用起来。\n\n\nZou, Ivan: 请问在build前端界面时,由于后端是使用的R和python,那么前端不直接用Python Flask或者R shiny构建的理由? 同样对于table和plot的display,有考虑直接使用Python或者R中的库去实现相同效果吗?\nZY: Javascript 做前端会更灵活和方便。互动性也会更好。Plotly 也有Python 版本吧。R 产生的图不支持互动。" } ] \ No newline at end of file diff --git a/docs/sitemap.xml b/docs/sitemap.xml index ca17a92..0fd603b 100644 --- a/docs/sitemap.xml +++ b/docs/sitemap.xml @@ -10,6 +10,10 @@ https://pharmarug.github.io/pharmarug-2024/program.html - 2024-03-29T16:01:04.169Z + 2024-04-08T03:00:45.714Z + + + https://pharmarug.github.io/pharmarug-2024/questions.html + 2024-04-08T02:57:07.762Z diff --git a/program.qmd b/program.qmd index 3e92884..5595c51 100644 --- a/program.qmd +++ b/program.qmd @@ -8,11 +8,11 @@ title: "China Pharma RUG Meeting Program" | --------------- | ------------------------------------------------------------------------------------------------------------- | ------------------------------------------ | ----------------- | | **8:30-9:00** | **Registration** | | | 9:00-9:10 | Welcome | Joe Zhu, Roche; Wang Zhang, MSD | Beijing, Shanghai | -| 9:10-9:15 | Opening ([video](resources/video/session1_1-2x-half.mp4)) Remarks | Yan Qiao, BeiGene | Beijing | +| 9:10-9:15 | Opening Remarks ([video](resources/video/session1_1-2x-half.mp4)) | Yan Qiao, BeiGene | Beijing | | 9:15-9:45 | [Dynamic CDISC Analysis Results Data with the R Packages {cards} and {gtsummary}](https://www.danieldsjoberg.com/china-pharma-keynote-2024/material.html) ([video](resources/video/session1_2-1dot5x-half.mp4)) | Daniel Sjoberg, Roche | Online | -| 9:45-10:10 | [Optimize decision-making efficiency and speed by performing exploratory analysis through the MedDRAH platform](resources/slides/session_1_2_zhiping_Optimize%20decision-making%20efficiency%20and%20speed%20by%20performing%20exploratory%20analysis%20through%20the%20MedDRAH%20platform-Zhiping%20Yan.pptx) ([video](resources/video/session1_3-1dot5x-half.mp4)) | Huadan Li, Zhiping Yan, Dizal | Beijing | +| 9:45-10:10 | [Optimize decision-making efficiency and speed by performing exploratory analysis through the MedDRAH platform](resources/slides/session_1_2_zhiping_Optimize%20decision-making%20efficiency%20and%20speed%20by%20performing%20exploratory%20analysis%20through%20the%20MedDRAH%20platform-Zhiping%20Yan.pptx) ([video](resources/video/session1_3-1dot5x-half.mp4) [question](questions.html#optimize-decision-making-efficiency-and-speed-by-performing-exploratory-analysis-through-the-meddrah-platform)) | Huadan Li, Zhiping Yan, Dizal | Beijing | | **10:10-10:40** | **Break** | | | -| 10:40-11:05 | [Leveraging R for Real-Time Data Analysis and Reporting in the AI+HI Paradigm](resources/slides/session_2_1_hao_Leveraging%20R%20for%20Real-Time%20Data%20Analysis%20and%20Reporting%20in%20the%20AI%2BHI%20Paradigm_HaoChen%20(1).pptx) ([video](resources/video/session2_1-1dot5x-half.mp4)) | Hao Chen, Beigene | Shanghai | +| 10:40-11:05 | [Leveraging R for Real-Time Data Analysis and Reporting in the AI+HI Paradigm](resources/slides/session_2_1_hao_Leveraging%20R%20for%20Real-Time%20Data%20Analysis%20and%20Reporting%20in%20the%20AI%2BHI%20Paradigm_HaoChen%20(1).pptx) ([video](resources/video/session2_1-1dot5x-half.mp4) [question](questions.html#leveraging-r-for-real-time-data-analysis-and-reporting-in-the-aihi-paradigm)) | Hao Chen, Beigene | Shanghai | | 11:05-11:30 | [From XPT to Dataset-JSON as Alternative Transport Format for Regulatory Submission in R](resources/slides/session_2_2_huan_XPT_to_Dataset-JSON.pptx) ([video](resources/video/session2_2-1dot5x-half.mp4)) | Huan Lu, Neo Yang, Sanofi | Beijing | | 11:30-11:55 | [Create TLGs and log files by sassy](resources/slides/session_2_3_perphy_pharmaR2024_Perphy.pdf) ([video](resources/video/session2_3-1dot5x-half.mp4)) | Perphy Zhao, Sanofi | Beijing | | **12:00-13:00** | **Lunch Break** | @@ -23,9 +23,9 @@ title: "China Pharma RUG Meeting Program" | **14:40-15:10** | **Break** | | | 15:10-15:35 | [Chevron: A open-source package to facilitate generation of outputs](resources/slides/session_4_1_liming_RUG%202024%20chevron%20updated.pdf) ([video](resources/video/session4_1-1dot5x-half.mp4)) | Liming Li, Roche | Shanghai | | 15:35-16:00 | [R Shiny Tool for Mayo Score Monitoring in UC Study](resources/slides/session_4_2_shaoming_PharmaRUG_Mayo%20Score%20Shiny%20App_final.pptx) [demo](resources/slides/session_4_2_shaoming/pharmarug_mayo_score-2x.mp4) ([video](resources/video/session4_2-1dot5x-half.mp4)) | Shaoming Liu, Pfizer | Shanghai | -| 16:00-16:25 | [DaVinci Journey in the Early Phase Oncology Study](resources/slides/session_4_3_PharmaRUG_DaVinci_Yifan_final.pdf) ([video 1](resources/video/session4_3_1-1dot5x-half.mp4)) ([video 2](resources/video/session4_3_2-1dot5x-half.mp4)) | Yifan Han, Boehringer Ingelheim | Shanghai | +| 16:00-16:25 | [DaVinci Journey in the Early Phase Oncology Study](resources/slides/session_4_3_PharmaRUG_DaVinci_Yifan_final.pdf) ([video 1](resources/video/session4_3_1-1dot5x-half.mp4) [video 2](resources/video/session4_3_2-1dot5x-half.mp4) [question](questions.html#davinci-journey-in-the-early-phase-oncology-study)) | Yifan Han, Boehringer Ingelheim | Shanghai | | 16:25-16:50 | [Applying R to Big Data Processing ——Dtplyr as an Example](resources/slides/session_4_4_kaiping_Applying+R+to+Big+Data+Processing/Applying-R-to-Big-Data-Processing.html) ([video](resources/video/session4_4-1dot5x-half.mp4)) | Kaiping Yang, Beigene | Online | -| 16:50-17:00 | Closing ([video](resources/video/session4_5-1dot5x-half.mp4))Remarks | Yanli Chang, Novartis | Shanghai | +| 16:50-17:00 | Closing Remarks ([video](resources/video/session4_5-1dot5x-half.mp4)) | Yanli Chang, Novartis | Shanghai | ### Dynamic CDISC Analysis Results Data with the R Packages {cards} and {gtsummary}