Papers
arxiv:2309.14568

Introducing DictaLM -- A Large Generative Language Model for Modern Hebrew

Published on Sep 25, 2023
Authors:
,
,

Abstract

DictaLM and DictaLM-Rab are large-scale language models for Modern Hebrew and Rabbinic/Historical Hebrew, respectively, designed for various Hebrew-specific tasks and released under a Creative Commons license.

AI-generated summary

We present DictaLM, a large-scale language model tailored for Modern Hebrew. Boasting 7B parameters, this model is predominantly trained on Hebrew-centric data. As a commitment to promoting research and development in the Hebrew language, we release both the foundation model and the instruct-tuned model under a Creative Commons license. Concurrently, we introduce DictaLM-Rab, another foundation model geared towards Rabbinic/Historical Hebrew. These foundation models serve as ideal starting points for fine-tuning various Hebrew-specific tasks, such as instruction, Q&A, sentiment analysis, and more. This release represents a preliminary step, offering an initial Hebrew LLM model for the Hebrew NLP community to experiment with.

Community

על מה מדובר? מי כתב אותך?

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2309.14568 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2309.14568 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2309.14568 in a Space README.md to link it from this page.

Collections including this paper 2