set default api base for ai models to localhost
This commit is contained in:
		
							parent
							
								
									0bd483147d
								
							
						
					
					
						commit
						8c36fe5a72
					
				
					 2 changed files with 5 additions and 13 deletions
				
			
		|  | @ -51,14 +51,6 @@ | ||||||
|     fprintd = { |     fprintd = { | ||||||
|       enable = true; |       enable = true; | ||||||
|     }; |     }; | ||||||
|     ollama = { |  | ||||||
|       enable = false; |  | ||||||
| 
 |  | ||||||
|       loadModels = [ |  | ||||||
|         "deepseek-coder:1.3b" |  | ||||||
|         "deepseek-r1:1.5b" |  | ||||||
|       ]; |  | ||||||
|     }; |  | ||||||
|     tailscale = { |     tailscale = { | ||||||
|       enable = true; |       enable = true; | ||||||
|       authKeyFile = config.sops.secrets."vpn-keys/tailscale-authkey/horizon".path; |       authKeyFile = config.sops.secrets."vpn-keys/tailscale-authkey/horizon".path; | ||||||
|  |  | ||||||
|  | @ -16,10 +16,10 @@ | ||||||
|               type = lib.types.str; |               type = lib.types.str; | ||||||
|               default = "ollama"; |               default = "ollama"; | ||||||
|             }; |             }; | ||||||
|             # apiBase = lib.mkOption { |             apiBase = lib.mkOption { | ||||||
|             #   type = lib.types.nullOr lib.types.str; |               type = lib.types.str; | ||||||
|             #   default = null; |               default = "http://localhost:11434"; | ||||||
|             # }; |             }; | ||||||
|             roles = lib.mkOption { |             roles = lib.mkOption { | ||||||
|               type = lib.types.listOf (lib.types.enum [ |               type = lib.types.listOf (lib.types.enum [ | ||||||
|                 "chat" |                 "chat" | ||||||
|  | @ -40,7 +40,7 @@ | ||||||
|   }; |   }; | ||||||
| 
 | 
 | ||||||
|   config = { |   config = { | ||||||
|     # TODO: configure ollama to download any modules listed in options.host.ai.models.{name}.model if options.host.ai.models.{name}.apiBase is null |     # TODO: configure ollama to download any modules listed in options.host.ai.models.{name}.model if options.host.ai.models.{name}.apiBase is the default value | ||||||
|     # TODO: if we have any models that have a non null options.host.ai.models.{name}.apiBase then set services.ollama.enable to a lib.mkAfter true |     # TODO: if we have any models that have a non null options.host.ai.models.{name}.apiBase then set services.ollama.enable to a lib.mkAfter true | ||||||
|   }; |   }; | ||||||
| } | } | ||||||
|  |  | ||||||
		Loading…
	
	Add table
		Add a link
		
	
		Reference in a new issue