Home
Categories
EXPLORE
Music
Comedy
Society & Culture
History
Education
Business
True Crime
About Us
Contact Us
Copyright
© 2024 PodJoint
00:00 / 00:00
Sign in

or

Don't have an account?
Sign up
Forgot password
https://is1-ssl.mzstatic.com/image/thumb/Podcasts116/v4/87/8b/1e/878b1e67-fd1a-fb2f-de5b-113fe4018dc7/mza_11173054665888442467.jpg/600x600bb.jpg
TechcraftingAI NLP
Brad Edwards
271 episodes
3 days ago
TechcraftingAI NLP brings you daily summaries of the latest arXiv Computation and Language research.
Show more...
Technology
RSS
All content for TechcraftingAI NLP is the property of Brad Edwards and is served directly from their servers with no modification, redirects, or rehosting. The podcast is not affiliated with or endorsed by Podjoint in any way.
TechcraftingAI NLP brings you daily summaries of the latest arXiv Computation and Language research.
Show more...
Technology
https://d3t3ozftmdmh3i.cloudfront.net/staging/podcast_uploaded_nologo/39368654/39368654-1703088924475-7aa75231d6474.jpg
Ep. 257 - June 7, 2024
TechcraftingAI NLP
52 minutes 15 seconds
1 year ago
Ep. 257 - June 7, 2024

ArXiv NLP research for Friday, June 07, 2024.


00:19: Key-Element-Informed sLLM Tuning for Document Summarization

01:22: Low-Resource Cross-Lingual Summarization through Few-Shot Learning with Large Language Models

02:42: Large Language Model-guided Document Selection

04:13: More Victories, Less Cooperation: Assessing Cicero's Diplomacy Play

05:24: DiNeR: a Large Realistic Dataset for Evaluating Compositional Generalization

06:43: MATTER: Memory-Augmented Transformer Using Heterogeneous Knowledge Sources

08:01: Mixture-of-Agents Enhances Large Language Model Capabilities

09:09: AICoderEval: Improving AI Domain Code Generation of Large Language Models

11:00: CRAG -- Comprehensive RAG Benchmark

13:04: CRiskEval: A Chinese Multi-Level Risk Evaluation Benchmark Dataset for Large Language Models

14:52: Think out Loud: Emotion Deducing Explanation in Dialogues

16:43: WildBench: Benchmarking LLMs with Challenging Tasks from Real Users in the Wild

18:46: SelfGoal: Your Language Agents Already Know How to Achieve High-level Goals

19:58: BERTs are Generative In-Context Learners

20:43: Annotating FrameNet via Structure-Conditioned Language Generation

21:49: Revisiting Catastrophic Forgetting in Large Language Model Tuning

22:43: FedLLM-Bench: Realistic Benchmarks for Federated Learning of Large Language Models

24:33: Do Language Models Exhibit Human-like Structural Priming Effects?

25:27: Uncertainty Aware Learning for Language Model Alignment

26:50: The Russian Legislative Corpus

27:24: ComplexTempQA: A Large-Scale Dataset for Complex Temporal Question Answering

28:53: HateDebias: On the Diversity and Variability of Hate Speech Debiasing

30:29: A Deep Dive into the Trade-Offs of Parameter-Efficient Preference Alignment Techniques

32:00: Sexism Detection on a Data Diet

33:18: XTTS: a Massively Multilingual Zero-Shot Text-to-Speech Model

34:21: Through the Thicket: A Study of Number-Oriented LLMs derived from Random Forest Models

35:32: LLM-based speaker diarization correction: A generalizable approach

36:52: TCMD: A Traditional Chinese Medicine QA Dataset for Evaluating Large Language Models

38:10: BAMO at SemEval-2024 Task 9: BRAINTEASER: A Novel Task Defying Common Sense

39:10: Quantifying Geospatial in the Common Crawl Corpus

40:14: MEFT: Memory-Efficient Fine-Tuning through Sparse Adapter

41:47: Language models emulate certain cognitive profiles: An investigation of how predictability measures interact with individual differences

43:19: Compositional Generalization with Grounded Language Models

44:26: Scenarios and Approaches for Situated Natural Language Explanations

46:04: Are Large Language Models More Empathetic than Humans?

47:38: SUMIE: A Synthetic Benchmark for Incremental Entity Summarization

48:52: Multi-Head RAG: Solving Multi-Aspect Problems with LLMs

50:33: An Empirical Study on Parameter-Efficient Fine-Tuning for MultiModal Large Language Models

TechcraftingAI NLP
TechcraftingAI NLP brings you daily summaries of the latest arXiv Computation and Language research.