mirror of https://github.com/ollama/ollama.git
				
				
				
			
		
			
				
	
	
		
			16 lines
		
	
	
		
			540 B
		
	
	
	
		
			Go
		
	
	
	
			
		
		
	
	
			16 lines
		
	
	
		
			540 B
		
	
	
	
		
			Go
		
	
	
	
| package llm
 | |
| 
 | |
| import (
 | |
| 	"fmt"
 | |
| 
 | |
| 	"github.com/jmorganca/ollama/api"
 | |
| )
 | |
| 
 | |
| func newDefaultExtServer(model string, adapters, projectors []string, opts api.Options) (extServer, error) {
 | |
| 	// On windows we always load the llama.cpp libraries dynamically to avoid startup DLL dependencies
 | |
| 	// This ensures we can update the PATH at runtime to get everything loaded
 | |
| 
 | |
| 	// This should never happen as we'll always try to load one or more cpu dynamic libaries before hitting default
 | |
| 	return nil, fmt.Errorf("no available default llm library on windows")
 | |
| }
 |