The paper introduces Guided Deep List, a tool for automating the generation of epidemiological line lists from open source reports. The tool uses distributed vector representations and dependency parsing to extract tabular data on disease outbreaks. It was evaluated on MERS outbreak data in Saudi Arabia, demonstrating improved accuracy over baseline methods and enabling epidemiological inferences.
This study investigates the correlation between Google Trends data for COVID-19 symptoms and the actual number of COVID-19 cases in Saudi Arabia between March and October 2020. The researchers found that searches for "cough" and "sore throat" were most frequent, while "loss of smell", "loss of taste", and "diarrhea" showed the highest correlation with confirmed cases. The study concludes that Google searches can serve as a supplementary surveillance tool for monitoring the spread of COVID-19 in Saudi Arabia. Why it matters: The research demonstrates the potential of using readily available digital data to augment traditional surveillance methods for public health monitoring in the region.
A new dataset called the Saudi Privacy Policy Dataset is introduced, which contains Arabic privacy policies from various sectors in Saudi Arabia. The dataset is annotated based on the 10 principles of the Personal Data Protection Law (PDPL) and includes 1,000 websites, 4,638 lines of text, and 775,370 tokens. The dataset aims to facilitate research and development in privacy policy analysis, NLP, and machine learning applications related to data protection.
A new methodology emulating fact-checker criteria assesses news outlet factuality and bias using LLMs. The approach uses prompts based on fact-checking criteria to elicit and aggregate LLM responses for predictions. Experiments demonstrate improvements over baselines, with error analysis on media popularity and region, and a released dataset/code at https://github.com/mbzuai-nlp/llm-media-profiling.
This paper introduces ProgramFC, a fact-checking model that decomposes complex claims into simpler sub-tasks using a library of functions. The model uses LLMs to generate reasoning programs and executes them by delegating sub-tasks, enhancing explainability and data efficiency. Experiments on fact-checking datasets demonstrate ProgramFC's superior performance compared to baseline methods, with publicly available code and data.