--- language: en license: mit library_name: transformers tags: - economics - finance - bert - language-model - financial-nlp - economic-analysis datasets: - custom_economic_corpus metrics: - accuracy - f1 - precision - recall pipeline_tag: fill-mask --- # EconBERT ## Model Description EconBERT is a BERT-based language model specifically fine-tuned for economic and financial text analysis. The model is designed to capture domain-specific language patterns, terminology, and contextual relationships in economic literature, research papers, financial reports, and related documents. > **Note**: The complete details of model architecture, training methodology, evaluation, and performance metrics are available in our paper. Please refer to the citation section below. ## Intended Uses & Limitations ### Intended Uses - **Economic Text Classification**: Categorizing economic documents, papers, or news articles - **Sentiment Analysis**: Analyzing market sentiment in financial news and reports - **Information Extraction**: Extracting structured data from unstructured economic texts - etc. ### Limitations - The model is specialized for economic and financial domains and may not perform as well on general text - Performance may vary on highly technical economic sub-domains not well-represented in the training data - For detailed discussion of limitations, please refer to our paper ## Training Data EconBERT was trained on a large corpus of economic and financial texts. For comprehensive information about the training data, including sources, size, preprocessing steps, and other details, please refer to our paper. ## Evaluation Results We evaluated EconBERT on several economic NLP tasks and compared its performance with general-purpose and other domain-specific models. The detailed evaluation methodology and complete results are available in our paper. Key findings include: - Improved performance on economic domain tasks compared to general BERT models - State-of-the-art results on [specific tasks, if applicable] - [Any other high-level results worth highlighting] ## How to Use ```python from transformers import AutoTokenizer, AutoModel # Load model and tokenizer tokenizer = AutoTokenizer.from_pretrained("YourUsername/EconBERT") model = AutoModel.from_pretrained("YourUsername/EconBERT") # Example usage text = "The Federal Reserve increased interest rates by 25 basis points." inputs = tokenizer(text, return_tensors="pt") outputs = model(**inputs) ``` For task-specific fine-tuning and applications, please refer to our paper and the examples provided in our GitHub repository. ## Citation If you use EconBERT in your research, please cite our paper: ```bibtex @article{LastName2025econbert, title={EconBERT: A Large Language Model for Economics}, author={Zhang, Philip and Rojcek, Jakub and Leippold, Markus}, journal={SSRN Working Paper}, year={2025}, volume={}, pages={}, publisher={University of Zurich}, doi={} } ``` ## Additional Information - **Model Type**: BERT - **Language(s)**: English - **License**: MIT For more detailed information about model architecture, training methodology, evaluation results, and applications, please refer to our paper.