Hyper-v

Bulk Skimming AI Paper Abstracts – Oct 9, 2024



Read/listen to the newsletter/podcast on substack:

My Twitter, LinkedIn, Discord, Patreon, consultation booking page, etc:

Timestamps:
00:00 intro
00:58 Scaling Proprioceptive-Visual Learning with Heterogeneous Pre-trained Transformers
01:48 COLLAGE – Collaborative Human-Agent Interaction Generation using Hierarchical Latent Diffusion and LMs
02:15 The Perfect Blend – Redefining RLHF with Mixture of Judges
04:38 A Looming Replication Crisis in Evaluating Behavior in LMs? Evidence and Solutions
05:18 1 Trillion Token (1TT) Platform – A Novel Framework for Efficient Data Sharing and Compensation in LLMs
06:27 Counter-Current Learning – A Biologically Plausible Dual Network Approach for Deep Learning
07:49 Unifying back-propagation and forward-forward algorithms through model predictive control
09:08 Can LLMs Really Learn to Translate a Low-Resource Language from One Grammar Book?
09:45 HM3 – Hierarchical Multi-Objective Model Merging for Pretrained Models
10:22 Hierarchical Federated ADMM
10:51 Cottention – Linear Transformers With Cosine Attention
15:18 Effects of AI Feedback on Learning, the Skill Gap, and Intellectual Diversity
18:57 Kolmogorov-Arnold Network AEs
20:12 PeerArg – Argumentative Peer Review with LLMs
21:55 When a LM is optimized for reasoning, does it still show embers of autoregression – An analysis of OpenAI o1
25:12 Adaptive Inference-Time Compute – LLMs Can Predict if They Can Do Better, Even Mid-Generation
26:29 LLMs Know More Than They Show – On the Intrinsic Representation of LLM Hallucinations
27:06 Selective Attention Improves Transformer
27:59 On the Proper Treatment of Tokenization in Psycholinguistics
28:32 FAN – Fourier Analysis Networks
30:25 Generalization emerges from local optimization in a self-organized learning network
32:35 Fair Decentralized Learning
33:18 Intelligence at the Edge of Chaos
35:11 Post-edits Are Preferences Too
36:06 Theoretical Insights into Fine-Tuning Attention – Generalization and Optimization
37:12 EmbedLLM – Learning Compact Representations of LLMs
37:53 Planning in Strawberry Fields – Evaluating and Improving the Planning and Scheduling Capabilities of LRM o1
38:33 Mitigating Memorization In LMs
39:22 U-shaped and Inverted-U Scaling behind Emergent Abilities of LLMs
40:36 ENTP – Encoder-only Next Token Prediction
42:07 House of Cards – Massive Weights in LLMs
44:16 Geometric Signatures of Compositionality Across a LM’s Lifetime
45:13 FlashMask – Efficient and Rich Mask Extension of FlashAttention
46:06 Sparse AEs Reveal Temporal Difference Learning in LLMs
46:37 nGPT – Normalized Transformer with Representation Learning on the Hypersphere
50:15 Draft on the Fly – Adaptive Self-Speculative Decoding using Cosine Similarity
50:49 Investigating the Synergistic Effects of Dropout and Residual Connections on LM Training
51:43 Do Music Generation Models Encode Music Theory?
52:08 RisingBALLER – A path towards a foundational model for football players data analytics
53:24 Self-Updatable LLMs with Parameter Integration
53:50 Stability analysis of chaotic systems in latent spaces
54:30 MoS – Unleashing Parameter Efficiency of LoRA with Mixture of Shards
55:51 Are LLMs Aware that Some Questions are not Open-ended?
56:56 TikGuard – A Deep Learning Transformer-Based Solution for Detecting Unsuitable TikTok Content for Kids
59:48 Vision LMs See What You Want but not What You See

[ad_2]

source

Related Articles

5 Comments

  1. imagine going back in time and explaining how i watched someone talk to a piece of metal about an autonomous heterogeneous piece of metal that got teleported to me using math about other math that someone else invented all while being on 2 different continents while talking to our mathematical algorithms and looking at a magical slab of glass that does math to show me what other people are doing , if that's not the future i dunno what is …😂😂😂

  2. Regarding the potential of increasing a skill gap: Since ChatGPT's release, it has become clear to me that individuals who employ critical thinking and creativity tend to get more out of language models. My hope is that LLMs will help people develop these skills. Rather than widening the gap, I believe this technology underscores the importance of helping others cultivate clear and structured thought processes.

Leave a Reply

Your email address will not be published. Required fields are marked *

Back to top button