--- library_name: peft license: apache-2.0 base_model: meta-llama/Llama-2-7b-hf tags: - resume-screening - hr-tech - llama2 - lora - peft - fine-tuned --- # Advanced Resume Screening Model ## Model Description This is a LoRA (Low-Rank Adaptation) fine-tuned version of Llama-2-7B specifically optimized for resume screening and candidate evaluation tasks. The model can analyze resumes, extract key information, and provide structured assessments of candidate qualifications. - **Developed by:** kiritps - **Model type:** Causal Language Model (LoRA Fine-tuned) - **Language(s):** English - **License:** Apache 2.0 - **Finetuned from model:** meta-llama/Llama-2-7b-hf ## Model Sources - **Repository:** https://huggingface.co/kiritps/Advanced-resume-screening ## Uses ### Direct Use This model is designed for HR professionals and recruitment systems to: - Analyze and screen resumes automatically - Extract key qualifications and skills - Provide structured candidate assessments - Filter candidates based on specific criteria - Generate summaries of candidate profiles ### Downstream Use The model can be integrated into: - Applicant Tracking Systems (ATS) - HR management platforms - Recruitment automation tools - Candidate matching systems ### Out-of-Scope Use - Should not be used as the sole decision-maker in hiring processes - Not intended for discriminatory screening based on protected characteristics - Not suitable for general-purpose text generation outside of resume/HR context ## How to Get Started with the Model from transformers import AutoTokenizer, AutoModelForCausalLM from peft import PeftModel Load base model and tokenizer base_model = AutoModelForCausalLM.from_pretrained("meta-llama/Llama-2-7b-hf") tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-2-7b-hf") Load LoRA adapter model = PeftModel.from_pretrained(base_model, "kiritps/Advanced-resume-screening") Example usage prompt = "Analyze this resume and provide key qualifications: [RESUME TEXT HERE]" inputs = tokenizer(prompt, return_tensors="pt") outputs = model.generate(**inputs, max_length=512, temperature=0.7) response = tokenizer.decode(outputs, skip_special_tokens=True) text ## Training Details ### Training Data The model was fine-tuned on a curated dataset of resume-response pairs, designed to teach the model how to: - Extract relevant information from resumes - Provide structured analysis of candidate qualifications - Generate appropriate screening responses ### Training Procedure #### Training Hyperparameters - **Training regime:** 4-bit quantization with bfloat16 mixed precision - **LoRA rank:** 64 - **LoRA alpha:** 16 - **Learning rate:** 2e-4 - **Batch size:** 4 - **Gradient accumulation steps:** 4 - **Training epochs:** Multiple checkpoints saved (3840, 4320, 4800, 5280, 5760 steps) #### Quantization Configuration - **Quantization method:** bitsandbytes - **Load in 4bit:** True - **Quantization type:** nf4 - **Double quantization:** True - **Compute dtype:** bfloat16 ## Bias, Risks, and Limitations ### Limitations - Model responses should be reviewed by human recruiters - May exhibit biases present in training data - Performance may vary across different industries or job types - Requires careful prompt engineering for optimal results ### Recommendations - Use as a screening aid, not a replacement for human judgment - Regularly audit outputs for potential bias - Combine with diverse evaluation methods - Ensure compliance with local employment laws and regulations ## Technical Specifications ### Model Architecture - **Parameter Count:** ~7B parameters (base) + LoRA adapters - **Quantization:** 4-bit NF4 quantization ### Compute Infrastructure #### Hardware - GPU training environment - Compatible with consumer and enterprise GPUs #### Software - **Framework:** PyTorch - **PEFT Version:** 0.6.2 - **Transformers:** Latest compatible version - **Quantization:** bitsandbytes ## Training Procedure The following `bitsandbytes` quantization config was used during training: - quant_method: bitsandbytes - load_in_8bit: False - load_in_4bit: True - llm_int8_threshold: 6.0 - llm_int8_skip_modules: None - llm_int8_enable_fp32_cpu_offload: False - llm_int8_has_fp16_weight: False - bnb_4bit_quant_type: nf4 - bnb_4bit_use_double_quant: True - bnb_4bit_compute_dtype: bfloat16 ### Framework Versions - PEFT 0.6.2 - Transformers (compatible version) - PyTorch (latest stable) - bitsandbytes (for quantization) ## Model Card Authors kiritps ## Model Card Contact For questions or issues regarding this model, please open an issue in the model repository.