Short context? why?
#1
by
						
d00mus
	
							
						- opened
							
					
Both the parent models have longer context, up to 128k but this one is only 32k. which is really disappointing..Is it possible to fix that?
This comment has been hidden
			
			
				
					
	
	
				
	
	
Just re-checked..Qwen says that both, QwQ and Coder-instruct have 128k context.. But you mentioned Just Coder(no instruct?).. Actually, "no instruct" also has 128k..So it's really should not be a problem..
@d00mus The problem has been solved. Thank you for your feedback. The context has been changed to 128K!
@mradermacher Now that the problem of the context length has been solved, could you please provide the quantized version of the model again? Thank you so much for your help!