Quantization
#16 opened 9 months ago
		by
		
				
							
						jntjdbhvebjynfbjdv
	
Add exported onnx model 'model_qint8_arm64.onnx'
#15 opened 9 months ago
		by
		
				
							
						tonatiuh-synopsys
	
Matryoshka embeddings
									1
	#14 opened 11 months ago
		by
		
				
							
						ghidav
	
model doesnt seem to support device_map="auto" for multi GPU .
									2
	#11 opened about 1 year ago
		by
		
				
							
						pulkitmehtametacube
	
How to change the batch size? or any tips to make the model faster?
									1
	#7 opened over 1 year ago
		by
		
				
							
						sdvfsfwfgfwf
	
Help with error message "Could not locate the configuration.py inside Alibaba-NLP/new-impl."
								1
#6 opened over 1 year ago
		by
		
				
							
						azhang01