syeda5688.github.io
Open in
urlscan Pro
2606:50c0:8000::153
Public Scan
URL:
https://syeda5688.github.io/
Submission: On June 11 via api from US — Scanned from DE
Submission: On June 11 via api from US — Scanned from DE
Form analysis
0 forms found in the DOMText Content
Toggle navigation * about (current) * blog * publications * cv * ctrl k * SYED A. RIZVI PhD student at Yale University. Yale University New Haven, CT I am a first-year PhD student in Computer Science at Yale University, advised by Prof. David van Dijk. I am also currently working as a Student Researcher at Google for Summer 2024 under the supervision of Dr. Bryan Perozzi. I work on building and applying Graph Neural Networks, Large Language Models, and Foundational models to large-scale biological data. Before joining Yale, I had the privilege of working as an undergraduate research student at the University of Houston, Rice University, and Houston Methodist. I received my bachelor’s degree in Computer Science from the University of Houston in 2022. Here is my Google Scholar and LinkedIn. NEWS May 02, 2024 Our paper Cell2Sentence: Teaching Large Language Models the Language of Biology has been accepted at ICML 2024. Congratulations to all of the co-authors! See the blog post here. Jan 17, 2024 Our paper BrainLM: A Foundational Model for Brain Activity Recordings has been accepted at ICLR 2024. Congratulations to all of the co-authors! LATEST POSTS Jun 09, 2024 Basics of Graph Neural Networks SELECTED PUBLICATIONS 1. BrainLM: A foundation model for brain activity recordings Josue Ortega Caro, Antonio Henrique Oliveira Fonseca, Syed A Rizvi, and 8 more authors In , 2024 Abs Bib HTML We introduce the Brain Language Model (BrainLM), a foundation model for brain activity dynamics trained on 6,700 hours of fMRI recordings. Utilizing self-supervised masked-prediction training, BrainLM demonstrates proficiency in both fine-tuning and zero-shot inference tasks. Fine-tuning allows for the accurate prediction of clinical variables like age, anxiety, and PTSD as well as forecasting of future brain states. Critically, the model generalizes well to entirely new external cohorts not seen during training. In zero-shot inference mode, BrainLM can identify intrinsic functional networks directly from raw fMRI data without any network-based supervision during training. The model also generates interpretable latent representations that reveal relationships between brain activity patterns and cognitive states. Overall, BrainLM offers a versatile and interpretable framework for elucidating the complex spatiotemporal dynamics of human brain activity. It serves as a powerful "lens" through which massive repositories of fMRI data can be analyzed in new ways, enabling more effective interpretation and utilization at scale. The work demonstrates the potential of foundation models to advance computational neuroscience research. @inproceedings{caro2023brainlm, title = {BrainLM: A foundation model for brain activity recordings}, author = {Caro, Josue Ortega and de Oliveira Fonseca, Antonio Henrique and Rizvi, Syed A and Rosati, Matteo and Averill, Christopher and Cross, James L and Mittal, Prateek and Zappala, Emanuele and Dhodapkar, Rahul Madhav and Abdallah, Chadi and others}, url = {https://openreview.net/pdf?id=RwI7ZEfR27}, journal = {ICLR 2024}, year = {2024}, } 2. Cell2sentence: Teaching large language models the language of biology Daniel Levine, Sacha Lévy, Syed Asad Rizvi, and 8 more authors ICML 2024, 2024 Abs Bib HTML We introduce Cell2Sentence (C2S), a novel method to directly adapt large language models to a biological context, specifically single-cell transcriptomics. By transforming gene expression data into ”cell sentences,” C2S bridges the gap between natural language processing and biology. We demonstrate cell sentences enable the finetuning of language models for diverse tasks in biology, including cell generation, complex celltype annotation, and direct data-driven text generation. Our experiments reveal that GPT-2, when fine-tuned with C2S, can generate biologically valid cells based on cell type inputs, and accurately predict cell types from cell sentences. This illustrates that language models, through C2S finetuning, can acquire a significant understanding of single-cell biology while maintaining robust text generation capabilities. C2S offers a flexible, accessible framework to integrate natural language processing with transcriptomics, utilizing existing models and libraries for a wide range of biological applications. @article{levine2023cell2sentence, title = {Cell2sentence: Teaching large language models the language of biology}, author = {Levine, Daniel and L{\'e}vy, Sacha and Rizvi, Syed Asad and Pallikkavaliyaveetil, Nazreen and Chen, Xingyu and Zhang, David and Ghadermarzi, Sina and Wu, Ruiming and Zheng, Zihe and Vrkic, Ivan and others}, url = {https://www.biorxiv.org/content/biorxiv/early/2024/02/15/2023.09.11.557287.full.pdf}, journal = {ICML 2024}, year = {2024}, } Feel to reach out via email or social profiles! © Copyright 2024 Syed A. Rizvi. Powered by Jekyll with al-folio theme. Hosted by GitHub Pages.