-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
long context #360
Comments
https://twitter.com/Yampeleg/status/1674430869828956161?s=20 points to simple non-linear modification for improved short-range interpolation, similar to what we discussed |
^
|
#516 (comment) 70B with 16K |
huggingface/text-generation-inference#529 TGI will soon have it too |
Things I’m Learning While Training SuperHOT | kaiokendev.github.io Reddit - Dive into anything TheBloke/Llama-2-70B-chat-GPTQ · Hugging Face NTK-Aware Scaled RoPE allows LLaMA models to have extended (8k+) context size without any fine-tuning and minimal perplexity degradation. : LocalLLaMA Llama 2 is here - get it on Hugging Face jquesnelle/scaled-rope (Experimental) Add support to NTK RoPE scaling by Panchovix · Pull Request #118 · turboderp/exllama Reddit - https://preview.redd.it/2qdj7itsb39b1.png?width=662&format=png&auto=webp&s=464052174151b6ae8b6a9ce42b8f1acc9acabd35 How Long Can Open-Source LLMs Truly Promise on Context Length? | LMSYS Org DachengLi1/LongChat: Official repository for LongChat and LongEval Extending context size via RoPE scaling · ggerganov/llama.cpp · Discussion #1965 NTK-Aware Scaled RoPE allows LLaMA models to have extended (8k+) context size without any fine-tuning and minimal perplexity degradation. : LocalLLaMA HuggingFace models have Summary post for higher context sizes for this week. For context up to 4096, NTK RoPE scaling is pretty viable. For context higher than that, keep using SuperHOT LoRA/Merges. : LocalLLaMA Output garbled in llama2 model · Issue #510 · vllm-project/vllm Stay on topic with Classifier-Free Guidance : LocalLLaMA Add Classifier-Free Guidance sampling · Issue #24536 · huggingface/transformers tau/scrolls · Datasets at Hugging Face Quantized LLama2 70B GPTQ 4-bit · Issue #516 · h2oai/h2ogpt Request: NTK rope support · Issue #479 · vllm-project/vllm Add support for LLaMA-2 by zhuohan123 · Pull Request #505 · vllm-project/vllm lmsys/longchat-13b-16k · Hugging Face [2302.05507] Long-Context Language Decision Transformers and Exponential Tilt for Interactive Text Environments LongChat/longeval at longeval · DachengLi1/LongChat Request: NTK rope support · Issue #479 · vllm-project/vllm How Long Can Open-Source LLMs Truly Promise on Context Length? | LMSYS Org RoPE scaling support? · Issue #464 · vllm-project/vllm [2212.10947] Parallel Context Windows for Large Language Models [2307.03172] Lost in the Middle: How Language Models Use Long Contexts pseudotensor/LongChat: Official repository for LongChat and LongEval openchat/openchat · Hugging Face How Long Can Open-Source LLMs Truly Promise on Context Length? | LMSYS Org [2306.05685] Judging LLM-as-a-judge with MT-Bench and Chatbot Arena Training: Data: |
Giraffe: Adventures in Expanding Context Lengths in LLMs https://normxu.github.io/Rethinking-Rotary-Position-Embedding/ Consider also llama-2 way. |
https://lmsys.org/blog/2023-06-29-longchat/
https://arxiv.org/abs/2305.07185
https://www.reddit.com/r/LocalLLaMA/comments/14fgjqj/a_simple_way_to_extending_context_to_8k/
epfml/landmark-attention#1
epfml/landmark-attention@main...eugenepentland:landmark-attention-qlora:main
https://huggingface.co/epfml/landmark-attention-llama7b-wdiff
https://huggingface.co/TheBloke/landmark-attention-llama7b-fp16
https://huggingface.co/deepset/xlm-roberta-large-squad2
https://huggingface.co/SajjadAyoubi/xlm-roberta-large-fa-qa
https://arxiv.org/abs/2306.02707
https://www.reddit.com/r/LocalLLaMA/comments/14em713/just_released_vllm_inference_library_that/
https://github.com/vllm-project/vllm
https://vllm.ai/
https://arxiv.org/abs//2306.15595
The text was updated successfully, but these errors were encountered: