Last Week in Medical AI: Top Research Papers/Models ๐ฅ (November 2 -November 9, 2024)
๐ Medical AI Paper of the Week: Exploring Large Language Models for Specialist-level Oncology Care
Medical LLM & Other Models: - GSCo: Generalist-Specialist AI Collaboration - PediatricsGPT: Chinese Pediatric Assistant - MEG: Knowledge-Enhanced Medical QA - AutoProteinEngine: Multimodal Protein LLM
Frameworks and Methodologies: - BrainSegFounder: 3D Neuroimage Analysis - PASSION: Sub-Saharan Dermatology Dataset - SAM for Lung X-ray Segmentation - Label Critic: Data-First Approach - Medprompt Runtime Strategies
Medical LLM Applications: - CataractBot: Patient Support System - CheX-GPT: X-ray Report Enhancement - CardioAI: Cancer Cardiotoxicity Monitor - HealthQ: Healthcare Conversation Chain - PRObot: Diabetic Retinopathy Assistant
Medical LLMs & Benchmarks: - MediQ: Clinical Reasoning Benchmark - Touchstone: Segmentation Evaluation - Medical LLM Adaptation Progress - Fine-Tuning Medical QA Strategies
AI in Healthcare Ethics: - Healthcare Robotics with LLMs - XAI in Clinical Practice - Precision Rehabilitation Framework - Multimodal AI Challenges
Now you can watch and listen to the latest Medical AI papers daily on our YouTube and Spotify channels as well!
Looks like @Meta thinks we forgot they created PyTorch, so now they've open-sourced Lingua, a powerful and flexible library for training and inferencing large language models.
Things that stand out:
- Architecture: Pure PyTorch nn.Module implementation for easy customization.
- Checkpointing: Uses the new PyTorch distributed saving method (.distcp format) for flexible model reloading across different GPU configurations.
- Configuration: Utilizes data classes and YAML files for intuitive setup and modification.
- Profiling: Integrates with xFormers' profiler for automatic MFU and HFU calculation, plus memory profiling.
- Slurm Integration: Includes stool.py for seamless job launching on Slurm clusters.
Hyperdimensional Computing + Neural Network, tell your friends. To my knowledge, this is a completely novel implementation of HDC+Neural Networks. It would be a direct competitor to Transformers. It is off the charts more computationally efficient than Transformers could ever hope to be (which is why I tested it in the first place). It is far more similar to biological processes. My testing so far shows that it works surprisingly well. One surprise so far from my testing, adding an Attention Mechanism to the model does nothing at all. Weirdest thing. Like 1% performance increase. I guess Attention Is Not All You Need?
@GuangyuRobert (Twitter Handle) from MIT has created Project Sid, which simulates over 1,000 autonomous AI agents collaborating in a Minecraft environment, operating for extended periods without human intervention. This simulation demonstrates unprecedented levels of agent interaction, decision-making, and societal development.
Agents operate independently for hours or days, showcasing advanced decision-making algorithms and goal-oriented behavior.
The simulation produced complex, emergent phenomena, including: - Economic systems with currency (gems) and trading - Cultural development and religious practices - Agents even understood bribing. Priests were moving the most gems to bribe people into following them! - Governmental structures and democratic processes
Project Sid addresses fundamental challenges in AI research: - Coherence: Maintaining consistent agent behavior over extended periods. - Multi-agent Collaboration: Enabling effective communication and coordination among numerous AI entities. - Long-term Progression: Developing agents capable of learning and evolving over time.
While Minecraft serves as the initial testbed, the underlying AI architecture is designed to be game-agnostic, suggesting potential applications in various digital environments and real-world simulations.
Imagine a policy being debated by the government and how it might affect society; Sid can simulate its impact!
Even if this remains just a game experiment, the project successfully manages 1,000+ agents simultaneously, a feat that requires robust distributed computing and efficient agent architecture.
reacted to MonsterMMORPG's
post with ๐4 months ago
First fully multi-GPU supporting and very advanced batch image captioner APP with Gradio interface published (as far as i know first)
Multi-GPU batch caption with JoyCaption. JoyCaption uses Meta-Llama-3.1โ8B and google/siglip-so400m-patch14โ384 and a fine tuned image captioning neural network.
JUST RELEASED: Fireplace 2 for Llama 3.1 8b Instruct!
Fireplace 2 is an 'expansion pack' of structured outputs you can request during your chat, using special request tokens to let Llama know you're looking for specific types of responses: Inline function calls SQL queries JSON objects Data visualization with matplotlib
Sparse MoE (SMoE) has an unavoidable drawback: the performance of SMoE heavily relies on the choice of hyper-parameters, such as the number of activated experts per token (top-k) and the number of experts.
Also, identifying the optimal hyper-parameter without a sufficient number of ablation studies is challenging. As the size of the models continues to grow, this limitation could result in a significant waste of computational resources, and in turn, could hinder the efficiency of training MoE-based models in practice.
(READ MORE โโโ) Now, our DynMoE addresses these challenges! ๐ DynMoE incorporates: (1) a novel gating method that enables each token to automatically determine the number of experts to activate.
(2) An adaptive process automatically adjusts the number of experts during training. Extensive numerical results across Vision, Language, and Vision-Language tasks demonstrate the effectiveness of our approach to achieve competitive performance compared to GMoE for vision and language tasks, and MoE-LLaVA for vision-language tasks, while maintaining efficiency by activating fewer parameters.