Date | Title (arxiv) | Keyword | Affiliation | Note | Conference |
---|---|---|---|---|---|
2024.03 | Dynamic Contexts for Generating Suggestion Questions in RAG Based Conversational Systems | Dynamic Contexts | University of Illinois, Chicago | Makrdown | WWW`2024 workshop PromptEng |
2024.02 | Hint-before-Solving Prompting: Guiding LLMs to Effectively Utilize Encoded Knowledge | HSP, Summarization (not mentioned though) | National University of Singapore | Markdown | |
2024.02 | A Human-Inspired Reading Agent with Gist Memory of Very Long Contexts | Gist Memory, ReadAgent | Google DeepMind | Markdown | |
2024.01 | Meta-Prompting: Enhancing Language Models with Task-Agnostic Scaffolding | Meta-Prompting | Stanford University, OpenAI | Markdown |
Date | Title(arxiv) | Keyword | Affiliation | Note | Conference |
---|---|---|---|---|---|
2024.03 | PERL: Parameter Efficient Reinforcement Learning from Human Feedback | PERL, LoRA, Reward Model | Google Research | Markdown | |
2024.02 | ResLoRA: Identity Residual Mapping in Low-Rank Adaption | ResLoRA, Merge | Microsoft | Blog | |
2024.02 | LoRA+: Efficient Low Rank Adaptation of Large Models | LoRA+ | UC Berkeley | Markdown | |
2023.04 | LLaMA-Adapter V2: Parameter-Efficient Visual Instruction Model | LLaMA, Visual Instruction | Shanghai Artificial Intelligence Laboratory | Blog |
Date | Title(arxiv) | Keyword | Affiliation | Note | Conference |
---|---|---|---|---|---|
2024.01 | SliceGPT: Compress Large Language Models by Deleting Rows and Columns | Model Compression | Microsoft Research | Markdown | |
2023.04 | Learning to Compress Prompts with Gist Tokens | Gist Token, Compression | Stanford University | Blog | NeurIPS 2023 |
2022.05 | Matryoshka Representation Learning | Matryoshka, Adaptive | Google Research | Markdown | NeurIPS 2022 |
Date | Title(arxiv) | Keyword | Affiliation | Note | Conference |
---|---|---|---|---|---|
2023.04 | Chameleon: Plug-and-Play Compositional Reasoning with Large Language Models | Tools, Chameleon | University of California, Los Angeles, Microsoft Research | Blog | NeurIPS 2023 |
Date | Title (arxiv) | Keyword | Affiliation | Note | Conference |
---|---|---|---|---|---|
2024.02 | The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits | 1-bit, BitNet b1.58 | Microsoft Research | Blog | |
2024.02 | LongRoPE: Extending LLM Context Window Beyond 2 Million Tokens | LongRoPE, 2048k tokens | Microsoft Research | Makrdown | |
2023.04 | ResiDual: Transformer with Dual Residual Connections | Layer Normalization | Microsoft Research | Blog | ICLR 2024 rejected |
Date | Title (arxiv) | Keyword | Affiliation | Note | Conference |
---|---|---|---|---|---|
2023.03 | Visual Instruction Tuning | LLaVA, Visual Instruction Tuning | Microsoft Research | Blog | NeurIPS 2023 oral |
2021.08 | BROS: A Pre-trained Language Model Focusing on Text and Layout for Better Key Information Extraction from Documents | BROS, AMLM, TMLM | NAVER CLOVA | Markdown | AAAI 2022 |