Japanese SimCSE Collection Tsukagoshi et al., Japanese SimCSE Technical Report, arXiv 2023. https://arxiv.org/abs/2310.19349 • 5 items • Updated 16 days ago • 2
youko Collection The youko model series are based on the llama3 series and have been continually pre-trained on Japanese-specific corpora. • 9 items • Updated Jul 25 • 1
Llama 3.1 GPTQ, AWQ, and BNB Quants Collection Optimised Quants for high-throughput deployments! Compatible with Transformers, TGI & VLLM 🤗 • 9 items • Updated Jul 24 • 47
LLM-jp: A Cross-organizational Project for the Research and Development of Fully Open Japanese LLMs Paper • 2407.03963 • Published Jul 4 • 15
Llama-3-ELYZA-JP Collection Llama-3 models augmented for Japanese usage • 6 items • Updated Aug 14 • 8
view article Article An Analysis of Chinese LLM Censorship and Bias with Qwen 2 Instruct By leonardlin • Jun 11 • 45
view article Article The Open Medical-LLM Leaderboard: Benchmarking Large Language Models in Healthcare Apr 19 • 99
LEIA: Facilitating Cross-Lingual Knowledge Transfer in Language Models with Entity-based Data Augmentation Paper • 2402.11485 • Published Feb 18 • 1
Construction of Domain-specified Japanese Large Language Model for Finance through Continual Pre-training Paper • 2404.10555 • Published Apr 16 • 2
Pretraining and Updating Language- and Domain-specific Large Language Model: A Case Study in Japanese Business Domain Paper • 2404.08262 • Published Apr 12 • 1
Continual Pre-Training for Cross-Lingual LLM Adaptation: Enhancing Japanese Language Capabilities Paper • 2404.17790 • Published Apr 27 • 5
JMedLoRA:Medical Domain Adaptation on Japanese Large Language Models using Instruction-tuning Paper • 2310.10083 • Published Oct 16, 2023 • 2
JaColBERT and Hard Negatives, Towards Better Japanese-First Embeddings for Retrieval: Early Technical Report Paper • 2312.16144 • Published Dec 26, 2023 • 3
Karasu Collection The models trained under our Karasu and Qarasu project • 9 items • Updated 3 days ago • 1
NTQ AI LM Collection A collection of finely tuned Language Models (LLMs) across diverse datasets. • 3 items • Updated Apr 24 • 1
ELYZA-japanese-CodeLlama-7b Collection CodeLlama models augmented for Japanese usage • 3 items • Updated Aug 14 • 2
ELYZA-japanese-Llama-2-13b Collection 13b Llama-2 models augmented for Japanese usage • 5 items • Updated Aug 14 • 5
ELYZA-japanese-Llama-2-7b Collection 7b Llama-2 models augmented for Japanese usage • 6 items • Updated Aug 14 • 4
nekomata Collection The nekomata model series are based on the Qwen series and have been continually pre-trained on Japanese-specific corpora. • 8 items • Updated Jul 25 • 5
Japanese Multimodal Models Collection Suite of multimodal models focusing on Japan/Japanese-related usage • 4 items • Updated Apr 8 • 7
Japanese Stable LM Collection Suite of LLMs focusing on Japanese usage • 15 items • Updated May 7 • 16
youri Collection The youri model series are based on the llama2 series and have been continually pre-trained on Japanese-specific corpora. • 6 items • Updated Jul 25 • 1
bilingual-gpt-neox-4b Collection The bilingual-gpt-neox-4b series are pre-trained from scratch on a mixture of Japanese and English corpora. • 5 items • Updated Jul 25 • 1
japanese-gpt-neox-3.6b Collection The japanese-gpt-neox-3.6b series are pre-trained from scratch on Japanese corpora. • 5 items • Updated Jul 25 • 2