tokyotech-llm/Llama-3.1-Swallow-70B-Instruct-v0.3 Text Generation • Updated 3 days ago • 3.79k • • 10
Building Instruction-Tuning Datasets from Human-Written Instructions with Open-Weight Large Language Models Paper • 2503.23714 • Published 5 days ago
Balancing Speed and Stability: The Trade-offs of FP8 vs. BF16 Training in LLMs Paper • 2411.08719 • Published Nov 10, 2024
Why We Build Local Large Language Models: An Observational Analysis from 35 Japanese and Multilingual LLMs Paper • 2412.14471 • Published Dec 19, 2024
Wider or Deeper? Scaling LLM Inference-Time Compute with Adaptive Branching Tree Search Paper • 2503.04412 • Published about 1 month ago • 1
Wider or Deeper? Scaling LLM Inference-Time Compute with Adaptive Branching Tree Search Paper • 2503.04412 • Published about 1 month ago • 1
Drop-Upcycling: Training Sparse Mixture of Experts with Partial Re-initialization Paper • 2502.19261 • Published Feb 26 • 7
Drop-Upcycling: Training Sparse Mixture of Experts with Partial Re-initialization Paper • 2502.19261 • Published Feb 26 • 7
ExaGPT: Example-Based Machine-Generated Text Detection for Human Interpretability Paper • 2502.11336 • Published Feb 17
TAID: Temporally Adaptive Interpolated Distillation for Efficient Knowledge Transfer in Language Models Paper • 2501.16937 • Published Jan 28 • 6
Continual Pre-Training for Cross-Lingual LLM Adaptation: Enhancing Japanese Language Capabilities Paper • 2404.17790 • Published Apr 27, 2024 • 5
Building a Large Japanese Web Corpus for Large Language Models Paper • 2404.17733 • Published Apr 27, 2024 • 4
Agent Skill Acquisition for Large Language Models via CycleQD Paper • 2410.14735 • Published Oct 16, 2024 • 2
LLM-jp: A Cross-organizational Project for the Research and Development of Fully Open Japanese LLMs Paper • 2407.03963 • Published Jul 4, 2024 • 19
Continual Pre-Training for Cross-Lingual LLM Adaptation: Enhancing Japanese Language Capabilities Paper • 2404.17790 • Published Apr 27, 2024 • 5