fix HPU could not handle float16 in attention.py.
#68
by
						
sywangyi
	
							
						- opened
							
					
No description provided.
			
			
				
					
	
	
				
	
	
@yao-matrix
sywangyi
	
				
		changed pull request status to
		open
			
mpt model need to be changed in Attention for the following error. https://huggingface.co/mosaicml/mpt-7b/blob/main/attention.py#L49
causal_mask = attn_weight.new_ones(s, s, dtype=torch.float16)
RuntimeError: float16/half is not supported on Gaudi.
