2024-02-08 06:24:29 +08:00 
										
									 
								 
							 
							
								
							 
							
								 
							 
							
							
								# OpenAI compatibility
 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
									
										
										
										
											2025-02-08 01:55:07 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								>  [!NOTE]
 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								>  OpenAI compatibility is experimental and is subject to major adjustments including breaking changes. For fully-featured access to the Ollama API, see the Ollama [Python library](https://github.com/ollama/ollama-python), [JavaScript library](https://github.com/ollama/ollama-js) and [REST API](https://github.com/ollama/ollama/blob/main/docs/api.md).
 
							 
						 
					
						
							
								
									
										
										
										
											2024-02-08 06:24:29 +08:00 
										
									 
								 
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
									
										
										
										
											2024-02-08 06:25:24 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								Ollama provides experimental compatibility with parts of the [OpenAI API ](https://platform.openai.com/docs/api-reference ) to help connect existing applications to Ollama.
							 
						 
					
						
							
								
									
										
										
										
											2024-02-08 06:24:29 +08:00 
										
									 
								 
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								## Usage
 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								### OpenAI Python library
 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								```python
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								from openai import OpenAI
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								client = OpenAI(
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    base_url='http://localhost:11434/v1/',
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    # required but ignored
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    api_key='ollama',
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								)
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								chat_completion = client.chat.completions.create(
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    messages=[
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        {
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								            'role': 'user',
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								            'content': 'Say this is a test',
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        }
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    ],
							 
						 
					
						
							
								
									
										
										
										
											2024-09-26 02:11:22 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								    model='llama3.2',
							 
						 
					
						
							
								
									
										
										
										
											2024-02-08 06:24:29 +08:00 
										
									 
								 
							 
							
								
							 
							
								 
							 
							
							
								)
							 
						 
					
						
							
								
									
										
										
										
											2024-08-02 06:48:44 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
									
										
										
										
											2024-08-03 04:05:57 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								response = client.chat.completions.create(
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    model="llava",
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    messages=[
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        {
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								            "role": "user",
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								            "content": [
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								                {"type": "text", "text": "What's in this image?"},
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								                {
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								                    "type": "image_url",
							 
						 
					
						
							
								
									
										
										
										
											2024-10-24 01:31:04 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								                    "image_url": "data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAG0AAABmCAYAAADBPx+VAAAACXBIWXMAAAsTAAALEwEAmpwYAAAAAXNSR0IArs4c6QAAAARnQU1BAACxjwv8YQUAAA3VSURBVHgB7Z27r0zdG8fX743i1bi1ikMoFMQloXRpKFFIqI7LH4BEQ+NWIkjQuSWCRIEoULk0gsK1kCBI0IhrQVT7tz/7zZo888yz1r7MnDl7z5xvsjkzs2fP3uu71nNfa7lkAsm7d++Sffv2JbNmzUqcc8m0adOSzZs3Z+/XES4ZckAWJEGWPiCxjsQNLWmQsWjRIpMseaxcuTKpG/7HP27I8P79e7dq1ars/yL4/v27S0ejqwv+cUOGEGGpKHR37tzJCEpHV9tnT58+dXXCJDdECBE2Ojrqjh071hpNECjx4cMHVycM1Uhbv359B2F79+51586daxN/+pyRkRFXKyRDAqxEp4yMlDDzXG1NPnnyJKkThoK0VFd1ELZu3TrzXKxKfW7dMBQ6bcuWLW2v0VlHjx41z717927ba22U9APcw7Nnz1oGEPeL3m3p2mTAYYnFmMOMXybPPXv2bNIPpFZr1NHn4HMw0KRBjg9NuRw95s8PEcz/6DZELQd/09C9QGq5RsmSRybqkwHGjh07OsJSsYYm3ijPpyHzoiacg35MLdDSIS/O1yM778jOTwYUkKNHWUzUWaOsylE00MyI0fcnOwIdjvtNdW/HZwNLGg+sR1kMepSNJXmIwxBZiG8tDTpEZzKg0GItNsosY8USkxDhD0Rinuiko2gfL/RbiD2LZAjU9zKQJj8RDR0vJBR1/Phx9+PHj9Z7REF4nTZkxzX4LCXHrV271qXkBAPGfP/atWvu/PnzHe4C97F48eIsRLZ9+3a3f/9+87dwP1JxaF7/3r17ba+5l4EcaVo0lj3SBq5kGTJSQmLWMjgYNei2GPT1MuMqGTDEFHzeQSP2wi/jGnkmPJ/nhccs44jvDAxpVcxnq0F6eT8h4ni/iIWpR5lPyA6ETkNXoSukvpJAD3AsXLiwpZs49+fPn5ke4j10TqYvegSfn0OnafC+Tv9ooA/JPkgQysqQNBzagXY55nO/oa1F7qvIPWkRL12WRpMWUvpVDYmxAPehxWSe8ZEXL20sadYIozfmNch4QJPAfeJgW3rNsnzphBKNJM2KKODo1rVOMRYik5ETy3ix4qWNI81qAAirizgMIc+yhTytx0JWZuNI03qsrgWlGtwjoS9XwgUhWGyhUaRZZQNNIEwCiXD16tXcAHUs79co0vSD8rrJCIW98pzvxpAWyyo3HYwqS0+H0BjStClcZJT5coMm6D2LOF8TolGJtK9fvyZpyiC5ePFi9nc/oJU4eiEP0jVoAnHa9wyJycITMP78+eMeP37sXrx44d6+fdt6f82aNdkx1pg9e3Zb5W+RSRE+n+VjksQWifvVaTKFhn5O8my63K8Qabdv33b379/PiAP//vuvW7BggZszZ072/+TJk91YgkafPn166zXB1rQHFvouAWHq9z3SEevSUerqCn2/dDCeta2jxYbr69evk4MHDyY7d+7MjhMnTiTPnz9Pfv/+nfQT2ggpO2dMF8cghuoM7Ygj5iWCqRlGFml0QC/ftGmTmzt3rmsaKDsgBSPh0/8yPeLLBihLkOKJc0jp8H8vUzcxIA1k6QJ/c78tWEyj5P3o4u9+jywNPdJi5rAH9x0KHcl4Hg570eQp3+vHXGyrmEeigzQsQsjavXt38ujRo44LQuDDhw+TW7duRS1HGgMxhNXHgflaNTOsHyKvHK5Ijo2jbFjJBQK9YwFd6RVMzfgRBmEfP37suBBm/p49e1qjEP2mwTViNRo0VJWH1deMXcNK08uUjVUu7s/zRaL+oLNxz1bpANco4npUgX4G2eFbpDFyQoQxojBCpEGSytmOH8qrH5Q9vuzD6ofQylkCUmh8DBAr+q8JCyVNtWQIidKQE9wNtLSQnS4jDSsxNHogzFuQBw4cyM61UKVsjfr3ooBkPSqqQHesUPWVtzi9/vQi1T+rJj7WiTz4Pt/l3LxUkr5P2VYZaZ4URpsE+st/dujQoaBBYokbrz/8TJNQYLSonrPS9kUaSkPeZyj1AWSj+d+VBoy1pIWVNed8P0Ll/ee5HdGRhrHhR5GGN0r4LGZBaj8oFDJitBTJzIZgFcmU0Y8ytWMZMzJOaXUSrUs5RxKnrxmbb5YXO9VGUhtpXldhEUogFr3IzIsvlpmdosVcGVGXFWp2oU9kLFL3dEkSz6NHEY1sjSRdIuDFWEhd8KxFqsRi1uM/nz9/zpxnwlESONdg6dKlbsaMGS4EHFHtjFIDHwKOo46l4TxSuxgDzi+rE2jg+BaFruOX4HXa0Nnf1lwAPufZeF8/r6zD97WK2qFnGjBxTw5qNGPxT+5T/r7/7RawFC3j4vTp09koCxkeHjqbHJqArmH5UrFKKksnxrK7FuRIs8STfBZv+luugXZ2pR/pP9Ois4z+TiMzUUkUjD0iEi1fzX8GmXyuxUBRcaUfykV0YZnlJGKQpOiGB76x5GeWkWWJc3mOrK6S7xdND+W5N6XyaRgtWJFe13GkaZnKOsYqGdOVVVbGupsyA/l7emTLHi7vwTdirNEt0qxnzAvBFcnQF16xh/TMpUuXHDowhlA9vQVraQhkudRdzOnK+04ZSP3DUhVSP61YsaLtd/ks7ZgtPcXqPqEafHkdqa84X6aCeL7YWlv6edGFHb+ZFICPlljHhg0bKuk0CSvVznWsotRu433alNdFrqG45ejoaPCaUkWERpLXjzFL2Rpllp7PJU2a/v7Ab8N05/9t27Z16KUqoFGsxnI9EosS2niSYg9SpU6B4JgTrvVW1flt1sT+0ADIJU2maXzcUTraGCRaL1Wp9rUMk16PMom8QhruxzvZIegJjFU7LLCePfS8uaQdPny4jTTL0dbee5mYokQsXTIWNY46kuMbnt8Kmec+LGWtOVIl9cT1rCB0V8WqkjAsRwta93TbwNYoGKsUSChN44lgBNCoHLHzquYKrU6qZ8lolCIN0Rh6cP0Q3U6I6IXILYOQI513hJaSKAorFpuHXJNfVlpRtmYBk1Su1obZr5dnKAO+L10Hrj3WZW+E3qh6IszE37F6EB+68mGpvKm4eb9bFrlzrok7fvr0Kfv727dvWRmdVTJHw0qiiCUSZ6wCK+7XL/AcsgNyL74DQQ730sv78Su7+t/A36MdY0sW5o40ahslXr58aZ5HtZB8GH64m9EmMZ7FpYw4T6QnrZfgenrhFxaSiSGXtPnz57e9TkNZLvTjeqhr734CNtrK41L40sUQckmj1lGKQ0rC37x544r8eNXRpnVE3ZZY7zXo8NomiO0ZUCj2uHz58rbXoZ6gc0uA+F6ZeKS/jhRDUq8MKrTho9fEkihMmhxtBI1DxKFY9XLpVcSkfoi8JGnToZO5sU5aiDQIW716ddt7ZLYtMQlhECdBGXZZMWldY5BHm5xgAroWj4C0hbYkSc/jBmggIrXJWlZM6pSETsEPGqZOndr2uuuR5rF169a2HoHPdurUKZM4CO1WTPqaDaAd+GFGKdIQkxAn9RuEWcTRyN2KSUgiSgF5aWzPTeA/lN5rZubMmR2bE4SIC4nJoltgAV/dVefZm72AtctUCJU2CMJ327hxY9t7EHbkyJFseq+EJSY16RPo3Dkq1kkr7+q0bNmyDuLQcZBEPYmHVdOBiJyIlrRDq41YPWfXOxUysi5fvtyaj+2BpcnsUV/oSoEMOk2CQGlr4ckhBwaetBhjCwH0ZHtJROPJkyc7UjcYLDjmrH7ADTEBXFfOYmB0k9oYBOjJ8b4aOYSe7QkKcYhFlq3QYLQhSidNmtS2RATwy8YOM3EQJsUjKiaWZ+vZToUQgzhkHXudb/PW5YMHD9yZM2faPsMwoc7RciYJXbGuBqJ1UIGKKLv915jsvgtJxCZDubdXr165mzdvtr1Hz5LONA8jrUwKPqsmVesKa49S3Q4WxmRPUEYdTjgiUcfUwLx589ySJUva3oMkP6IYddq
							 
						 
					
						
							
								
									
										
										
										
											2024-08-03 04:05:57 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								                },
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								            ],
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        }
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    ],
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    max_tokens=300,
							 
						 
					
						
							
								
									
										
										
										
											2024-08-03 04:16:23 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								)
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								completion = client.completions.create(
							 
						 
					
						
							
								
									
										
										
										
											2024-09-26 02:11:22 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								    model="llama3.2",
							 
						 
					
						
							
								
									
										
										
										
											2024-08-03 04:16:23 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								    prompt="Say this is a test",
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								)
							 
						 
					
						
							
								
									
										
										
										
											2024-08-03 04:05:57 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
									
										
										
										
											2024-08-02 06:48:44 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								list_completion = client.models.list()
							 
						 
					
						
							
								
									
										
										
										
											2024-08-02 06:58:13 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
									
										
										
										
											2024-09-26 02:11:22 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								model = client.models.retrieve("llama3.2")
							 
						 
					
						
							
								
									
										
										
										
											2024-08-02 07:00:29 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								embeddings = client.embeddings.create(
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    model="all-minilm",
							 
						 
					
						
							
								
									
										
										
										
											2024-08-03 04:16:23 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								    input=["why is the sky blue?", "why is the grass green?"],
							 
						 
					
						
							
								
									
										
										
										
											2024-08-02 07:00:29 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								)
							 
						 
					
						
							
								
									
										
										
										
											2024-02-08 06:24:29 +08:00 
										
									 
								 
							 
							
								
							 
							
								 
							 
							
							
								```
							 
						 
					
						
							
								
									
										
										
										
											2025-02-08 01:55:07 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
									
										
										
										
											2024-12-07 02:35:37 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								#### Structured outputs
 
							 
						 
					
						
							
								
									
										
										
										
											2025-02-08 01:55:07 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								```python
							 
						 
					
						
							
								
									
										
										
										
											2024-12-08 14:14:36 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								from pydantic import BaseModel
							 
						 
					
						
							
								
									
										
										
										
											2024-12-07 02:35:37 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								from openai import OpenAI
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								client = OpenAI(base_url="http://localhost:11434/v1", api_key="ollama")
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								# Define the schema for the response
 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								class FriendInfo(BaseModel):
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    name: str
							 
						 
					
						
							
								
									
										
										
										
											2025-07-29 02:19:13 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								    age: int
							 
						 
					
						
							
								
									
										
										
										
											2024-12-07 02:35:37 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								    is_available: bool
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								class FriendList(BaseModel):
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    friends: list[FriendInfo]
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								try:
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    completion = client.beta.chat.completions.parse(
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        temperature=0,
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        model="llama3.1:8b",
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        messages=[
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								            {"role": "user", "content": "I have two friends. The first is Ollama 22 years old busy saving the world, and the second is Alonso 23 years old and wants to hang out. Return a list of friends in JSON format"}
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        ],
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        response_format=FriendList,
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    )
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    friends_response = completion.choices[0].message
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    if friends_response.parsed:
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        print(friends_response.parsed)
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    elif friends_response.refusal:
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        print(friends_response.refusal)
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								except Exception as e:
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    print(f"Error: {e}")
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								```
							 
						 
					
						
							
								
									
										
										
										
											2024-02-08 06:24:29 +08:00 
										
									 
								 
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								### OpenAI JavaScript library
 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								```javascript
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								import OpenAI from 'openai'
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								const openai = new OpenAI({
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								  baseURL: 'http://localhost:11434/v1/',
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								  // required but ignored
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								  apiKey: 'ollama',
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								})
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								const chatCompletion = await openai.chat.completions.create({
							 
						 
					
						
							
								
									
										
										
										
											2024-08-03 04:05:57 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								    messages: [{ role: 'user', content: 'Say this is a test' }],
							 
						 
					
						
							
								
									
										
										
										
											2024-09-26 02:11:22 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								    model: 'llama3.2',
							 
						 
					
						
							
								
									
										
										
										
											2024-08-03 04:05:57 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								})
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								const response = await openai.chat.completions.create({
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    model: "llava",
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    messages: [
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        {
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        role: "user",
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        content: [
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								            { type: "text", text: "What's in this image?" },
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								            {
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								            type: "image_url",
							 
						 
					
						
							
								
									
										
										
										
											2024-10-24 01:31:04 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								            image_url: "data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAG0AAABmCAYAAADBPx+VAAAACXBIWXMAAAsTAAALEwEAmpwYAAAAAXNSR0IArs4c6QAAAARnQU1BAACxjwv8YQUAAA3VSURBVHgB7Z27r0zdG8fX743i1bi1ikMoFMQloXRpKFFIqI7LH4BEQ+NWIkjQuSWCRIEoULk0gsK1kCBI0IhrQVT7tz/7zZo888yz1r7MnDl7z5xvsjkzs2fP3uu71nNfa7lkAsm7d++Sffv2JbNmzUqcc8m0adOSzZs3Z+/XES4ZckAWJEGWPiCxjsQNLWmQsWjRIpMseaxcuTKpG/7HP27I8P79e7dq1ars/yL4/v27S0ejqwv+cUOGEGGpKHR37tzJCEpHV9tnT58+dXXCJDdECBE2Ojrqjh071hpNECjx4cMHVycM1Uhbv359B2F79+51586daxN/+pyRkRFXKyRDAqxEp4yMlDDzXG1NPnnyJKkThoK0VFd1ELZu3TrzXKxKfW7dMBQ6bcuWLW2v0VlHjx41z717927ba22U9APcw7Nnz1oGEPeL3m3p2mTAYYnFmMOMXybPPXv2bNIPpFZr1NHn4HMw0KRBjg9NuRw95s8PEcz/6DZELQd/09C9QGq5RsmSRybqkwHGjh07OsJSsYYm3ijPpyHzoiacg35MLdDSIS/O1yM778jOTwYUkKNHWUzUWaOsylE00MyI0fcnOwIdjvtNdW/HZwNLGg+sR1kMepSNJXmIwxBZiG8tDTpEZzKg0GItNsosY8USkxDhD0Rinuiko2gfL/RbiD2LZAjU9zKQJj8RDR0vJBR1/Phx9+PHj9Z7REF4nTZkxzX4LCXHrV271qXkBAPGfP/atWvu/PnzHe4C97F48eIsRLZ9+3a3f/9+87dwP1JxaF7/3r17ba+5l4EcaVo0lj3SBq5kGTJSQmLWMjgYNei2GPT1MuMqGTDEFHzeQSP2wi/jGnkmPJ/nhccs44jvDAxpVcxnq0F6eT8h4ni/iIWpR5lPyA6ETkNXoSukvpJAD3AsXLiwpZs49+fPn5ke4j10TqYvegSfn0OnafC+Tv9ooA/JPkgQysqQNBzagXY55nO/oa1F7qvIPWkRL12WRpMWUvpVDYmxAPehxWSe8ZEXL20sadYIozfmNch4QJPAfeJgW3rNsnzphBKNJM2KKODo1rVOMRYik5ETy3ix4qWNI81qAAirizgMIc+yhTytx0JWZuNI03qsrgWlGtwjoS9XwgUhWGyhUaRZZQNNIEwCiXD16tXcAHUs79co0vSD8rrJCIW98pzvxpAWyyo3HYwqS0+H0BjStClcZJT5coMm6D2LOF8TolGJtK9fvyZpyiC5ePFi9nc/oJU4eiEP0jVoAnHa9wyJycITMP78+eMeP37sXrx44d6+fdt6f82aNdkx1pg9e3Zb5W+RSRE+n+VjksQWifvVaTKFhn5O8my63K8Qabdv33b379/PiAP//vuvW7BggZszZ072/+TJk91YgkafPn166zXB1rQHFvouAWHq9z3SEevSUerqCn2/dDCeta2jxYbr69evk4MHDyY7d+7MjhMnTiTPnz9Pfv/+nfQT2ggpO2dMF8cghuoM7Ygj5iWCqRlGFml0QC/ftGmTmzt3rmsaKDsgBSPh0/8yPeLLBihLkOKJc0jp8H8vUzcxIA1k6QJ/c78tWEyj5P3o4u9+jywNPdJi5rAH9x0KHcl4Hg570eQp3+vHXGyrmEeigzQsQsjavXt38ujRo44LQuDDhw+TW7duRS1HGgMxhNXHgflaNTOsHyKvHK5Ijo2jbFjJBQK9YwFd6RVMzfgRBmEfP37suBBm/p49e1qjEP2mwTViNRo0VJWH1deMXcNK08uUjVUu7s/zRaL+oLNxz1bpANco4npUgX4G2eFbpDFyQoQxojBCpEGSytmOH8qrH5Q9vuzD6ofQylkCUmh8DBAr+q8JCyVNtWQIidKQE9wNtLSQnS4jDSsxNHogzFuQBw4cyM61UKVsjfr3ooBkPSqqQHesUPWVtzi9/vQi1T+rJj7WiTz4Pt/l3LxUkr5P2VYZaZ4URpsE+st/dujQoaBBYokbrz/8TJNQYLSonrPS9kUaSkPeZyj1AWSj+d+VBoy1pIWVNed8P0Ll/ee5HdGRhrHhR5GGN0r4LGZBaj8oFDJitBTJzIZgFcmU0Y8ytWMZMzJOaXUSrUs5RxKnrxmbb5YXO9VGUhtpXldhEUogFr3IzIsvlpmdosVcGVGXFWp2oU9kLFL3dEkSz6NHEY1sjSRdIuDFWEhd8KxFqsRi1uM/nz9/zpxnwlESONdg6dKlbsaMGS4EHFHtjFIDHwKOo46l4TxSuxgDzi+rE2jg+BaFruOX4HXa0Nnf1lwAPufZeF8/r6zD97WK2qFnGjBxTw5qNGPxT+5T/r7/7RawFC3j4vTp09koCxkeHjqbHJqArmH5UrFKKksnxrK7FuRIs8STfBZv+luugXZ2pR/pP9Ois4z+TiMzUUkUjD0iEi1fzX8GmXyuxUBRcaUfykV0YZnlJGKQpOiGB76x5GeWkWWJc3mOrK6S7xdND+W5N6XyaRgtWJFe13GkaZnKOsYqGdOVVVbGupsyA/l7emTLHi7vwTdirNEt0qxnzAvBFcnQF16xh/TMpUuXHDowhlA9vQVraQhkudRdzOnK+04ZSP3DUhVSP61YsaLtd/ks7ZgtPcXqPqEafHkdqa84X6aCeL7YWlv6edGFHb+ZFICPlljHhg0bKuk0CSvVznWsotRu433alNdFrqG45ejoaPCaUkWERpLXjzFL2Rpllp7PJU2a/v7Ab8N05/9t27Z16KUqoFGsxnI9EosS2niSYg9SpU6B4JgTrvVW1flt1sT+0ADIJU2maXzcUTraGCRaL1Wp9rUMk16PMom8QhruxzvZIegJjFU7LLCePfS8uaQdPny4jTTL0dbee5mYokQsXTIWNY46kuMbnt8Kmec+LGWtOVIl9cT1rCB0V8WqkjAsRwta93TbwNYoGKsUSChN44lgBNCoHLHzquYKrU6qZ8lolCIN0Rh6cP0Q3U6I6IXILYOQI513hJaSKAorFpuHXJNfVlpRtmYBk1Su1obZr5dnKAO+L10Hrj3WZW+E3qh6IszE37F6EB+68mGpvKm4eb9bFrlzrok7fvr0Kfv727dvWRmdVTJHw0qiiCUSZ6wCK+7XL/AcsgNyL74DQQ730sv78Su7+t/A36MdY0sW5o40ahslXr58aZ5HtZB8GH64m9EmMZ7FpYw4T6QnrZfgenrhFxaSiSGXtPnz57e9TkNZLvTjeqhr734CNtrK41L40sUQckmj1lGKQ0rC37x544r8eNXRpnVE3ZZY7zXo8NomiO0ZUCj2uHz58rbXoZ6gc0uA+F6ZeKS/jhRDUq8MKrTho9fEkihMmhxtBI1DxKFY9XLpVcSkfoi8JGnToZO5sU5aiDQIW716ddt7ZLYtMQlhECdBGXZZMWldY5BHm5xgAroWj4C0hbYkSc/jBmggIrXJWlZM6pSETsEPGqZOndr2uuuR5rF169a2HoHPdurUKZM4CO1WTPqaDaAd+GFGKdIQkxAn9RuEWcTRyN2KSUgiSgF5aWzPTeA/lN5rZubMmR2bE4SIC4nJoltgAV/dVefZm72AtctUCJU2CMJ327hxY9t7EHbkyJFseq+EJSY16RPo3Dkq1kkr7+q0bNmyDuLQcZBEPYmHVdOBiJyIlrRDq41YPWfXOxUysi5fvtyaj+2BpcnsUV/oSoEMOk2CQGlr4ckhBwaetBhjCwH0ZHtJROPJkyc7UjcYLDjmrH7ADTEBXFfOYmB0k9oYBOjJ8b4aOYSe7QkKcYhFlq3QYLQhSidNmtS2RATwy8YOM3EQJsUjKiaWZ+vZToUQgzhkHXudb/PW5YMHD9yZM2faPsMwoc7RciYJXbGuBqJ1UIGKKLv915jsvgtJxCZDubdXr165mzdvtr1Hz5LONA8jrUwKPqsmVesKa49S3Q4WxmRPUEYdTjgiUcfUwLx589ySJUva3oMkP6IYddq6HMS4o55xB
							 
						 
					
						
							
								
									
										
										
										
											2024-08-03 04:05:57 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								            },
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        ],
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        },
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    ],
							 
						 
					
						
							
								
									
										
										
										
											2024-02-08 06:24:29 +08:00 
										
									 
								 
							 
							
								
							 
							
								 
							 
							
							
								})
							 
						 
					
						
							
								
									
										
										
										
											2024-08-02 06:48:44 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
									
										
										
										
											2024-08-03 04:16:23 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								const completion = await openai.completions.create({
							 
						 
					
						
							
								
									
										
										
										
											2024-09-26 02:11:22 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								    model: "llama3.2",
							 
						 
					
						
							
								
									
										
										
										
											2024-08-03 04:16:23 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								    prompt: "Say this is a test.",
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								})
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
									
										
										
										
											2024-08-02 06:48:44 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								const listCompletion = await openai.models.list()
							 
						 
					
						
							
								
									
										
										
										
											2024-08-02 06:58:13 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
									
										
										
										
											2024-09-26 02:11:22 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								const model = await openai.models.retrieve("llama3.2")
							 
						 
					
						
							
								
									
										
										
										
											2024-08-02 07:00:29 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								const embedding = await openai.embeddings.create({
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								  model: "all-minilm",
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								  input: ["why is the sky blue?", "why is the grass green?"],
							 
						 
					
						
							
								
									
										
										
										
											2024-08-03 04:16:23 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								})
							 
						 
					
						
							
								
									
										
										
										
											2024-02-08 06:24:29 +08:00 
										
									 
								 
							 
							
								
							 
							
								 
							 
							
							
								```
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								### `curl`
 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
									
										
										
										
											2025-02-08 01:55:07 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								```shell
							 
						 
					
						
							
								
									
										
										
										
											2024-02-08 06:24:29 +08:00 
										
									 
								 
							 
							
								
							 
							
								 
							 
							
							
								curl http://localhost:11434/v1/chat/completions \
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    -H "Content-Type: application/json" \
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    -d '{
							 
						 
					
						
							
								
									
										
										
										
											2024-09-26 02:11:22 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								        "model": "llama3.2",
							 
						 
					
						
							
								
									
										
										
										
											2024-02-08 06:24:29 +08:00 
										
									 
								 
							 
							
								
							 
							
								 
							 
							
							
								        "messages": [
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								            {
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								                "role": "system",
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								                "content": "You are a helpful assistant."
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								            },
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								            {
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								                "role": "user",
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								                "content": "Hello!"
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								            }
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        ]
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    }'
							 
						 
					
						
							
								
									
										
										
										
											2024-07-03 02:50:56 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
									
										
										
										
											2024-08-03 04:05:57 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								curl http://localhost:11434/v1/chat/completions \
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								  -H "Content-Type: application/json" \
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								  -d '{
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    "model": "llava",
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    "messages": [
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								      {
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        "role": "user",
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        "content": [
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								          {
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								            "type": "text",
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								            "text": "What'\''s in this image?"
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								          },
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								          {
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								            "type": "image_url",
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								            "image_url": {
							 
						 
					
						
							
								
									
										
										
										
											2024-10-24 01:31:04 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								               "url": "data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAAG0AAABmCAYAAADBPx+VAAAACXBIWXMAAAsTAAALEwEAmpwYAAAAAXNSR0IArs4c6QAAAARnQU1BAACxjwv8YQUAAA3VSURBVHgB7Z27r0zdG8fX743i1bi1ikMoFMQloXRpKFFIqI7LH4BEQ+NWIkjQuSWCRIEoULk0gsK1kCBI0IhrQVT7tz/7zZo888yz1r7MnDl7z5xvsjkzs2fP3uu71nNfa7lkAsm7d++Sffv2JbNmzUqcc8m0adOSzZs3Z+/XES4ZckAWJEGWPiCxjsQNLWmQsWjRIpMseaxcuTKpG/7HP27I8P79e7dq1ars/yL4/v27S0ejqwv+cUOGEGGpKHR37tzJCEpHV9tnT58+dXXCJDdECBE2Ojrqjh071hpNECjx4cMHVycM1Uhbv359B2F79+51586daxN/+pyRkRFXKyRDAqxEp4yMlDDzXG1NPnnyJKkThoK0VFd1ELZu3TrzXKxKfW7dMBQ6bcuWLW2v0VlHjx41z717927ba22U9APcw7Nnz1oGEPeL3m3p2mTAYYnFmMOMXybPPXv2bNIPpFZr1NHn4HMw0KRBjg9NuRw95s8PEcz/6DZELQd/09C9QGq5RsmSRybqkwHGjh07OsJSsYYm3ijPpyHzoiacg35MLdDSIS/O1yM778jOTwYUkKNHWUzUWaOsylE00MyI0fcnOwIdjvtNdW/HZwNLGg+sR1kMepSNJXmIwxBZiG8tDTpEZzKg0GItNsosY8USkxDhD0Rinuiko2gfL/RbiD2LZAjU9zKQJj8RDR0vJBR1/Phx9+PHj9Z7REF4nTZkxzX4LCXHrV271qXkBAPGfP/atWvu/PnzHe4C97F48eIsRLZ9+3a3f/9+87dwP1JxaF7/3r17ba+5l4EcaVo0lj3SBq5kGTJSQmLWMjgYNei2GPT1MuMqGTDEFHzeQSP2wi/jGnkmPJ/nhccs44jvDAxpVcxnq0F6eT8h4ni/iIWpR5lPyA6ETkNXoSukvpJAD3AsXLiwpZs49+fPn5ke4j10TqYvegSfn0OnafC+Tv9ooA/JPkgQysqQNBzagXY55nO/oa1F7qvIPWkRL12WRpMWUvpVDYmxAPehxWSe8ZEXL20sadYIozfmNch4QJPAfeJgW3rNsnzphBKNJM2KKODo1rVOMRYik5ETy3ix4qWNI81qAAirizgMIc+yhTytx0JWZuNI03qsrgWlGtwjoS9XwgUhWGyhUaRZZQNNIEwCiXD16tXcAHUs79co0vSD8rrJCIW98pzvxpAWyyo3HYwqS0+H0BjStClcZJT5coMm6D2LOF8TolGJtK9fvyZpyiC5ePFi9nc/oJU4eiEP0jVoAnHa9wyJycITMP78+eMeP37sXrx44d6+fdt6f82aNdkx1pg9e3Zb5W+RSRE+n+VjksQWifvVaTKFhn5O8my63K8Qabdv33b379/PiAP//vuvW7BggZszZ072/+TJk91YgkafPn166zXB1rQHFvouAWHq9z3SEevSUerqCn2/dDCeta2jxYbr69evk4MHDyY7d+7MjhMnTiTPnz9Pfv/+nfQT2ggpO2dMF8cghuoM7Ygj5iWCqRlGFml0QC/ftGmTmzt3rmsaKDsgBSPh0/8yPeLLBihLkOKJc0jp8H8vUzcxIA1k6QJ/c78tWEyj5P3o4u9+jywNPdJi5rAH9x0KHcl4Hg570eQp3+vHXGyrmEeigzQsQsjavXt38ujRo44LQuDDhw+TW7duRS1HGgMxhNXHgflaNTOsHyKvHK5Ijo2jbFjJBQK9YwFd6RVMzfgRBmEfP37suBBm/p49e1qjEP2mwTViNRo0VJWH1deMXcNK08uUjVUu7s/zRaL+oLNxz1bpANco4npUgX4G2eFbpDFyQoQxojBCpEGSytmOH8qrH5Q9vuzD6ofQylkCUmh8DBAr+q8JCyVNtWQIidKQE9wNtLSQnS4jDSsxNHogzFuQBw4cyM61UKVsjfr3ooBkPSqqQHesUPWVtzi9/vQi1T+rJj7WiTz4Pt/l3LxUkr5P2VYZaZ4URpsE+st/dujQoaBBYokbrz/8TJNQYLSonrPS9kUaSkPeZyj1AWSj+d+VBoy1pIWVNed8P0Ll/ee5HdGRhrHhR5GGN0r4LGZBaj8oFDJitBTJzIZgFcmU0Y8ytWMZMzJOaXUSrUs5RxKnrxmbb5YXO9VGUhtpXldhEUogFr3IzIsvlpmdosVcGVGXFWp2oU9kLFL3dEkSz6NHEY1sjSRdIuDFWEhd8KxFqsRi1uM/nz9/zpxnwlESONdg6dKlbsaMGS4EHFHtjFIDHwKOo46l4TxSuxgDzi+rE2jg+BaFruOX4HXa0Nnf1lwAPufZeF8/r6zD97WK2qFnGjBxTw5qNGPxT+5T/r7/7RawFC3j4vTp09koCxkeHjqbHJqArmH5UrFKKksnxrK7FuRIs8STfBZv+luugXZ2pR/pP9Ois4z+TiMzUUkUjD0iEi1fzX8GmXyuxUBRcaUfykV0YZnlJGKQpOiGB76x5GeWkWWJc3mOrK6S7xdND+W5N6XyaRgtWJFe13GkaZnKOsYqGdOVVVbGupsyA/l7emTLHi7vwTdirNEt0qxnzAvBFcnQF16xh/TMpUuXHDowhlA9vQVraQhkudRdzOnK+04ZSP3DUhVSP61YsaLtd/ks7ZgtPcXqPqEafHkdqa84X6aCeL7YWlv6edGFHb+ZFICPlljHhg0bKuk0CSvVznWsotRu433alNdFrqG45ejoaPCaUkWERpLXjzFL2Rpllp7PJU2a/v7Ab8N05/9t27Z16KUqoFGsxnI9EosS2niSYg9SpU6B4JgTrvVW1flt1sT+0ADIJU2maXzcUTraGCRaL1Wp9rUMk16PMom8QhruxzvZIegJjFU7LLCePfS8uaQdPny4jTTL0dbee5mYokQsXTIWNY46kuMbnt8Kmec+LGWtOVIl9cT1rCB0V8WqkjAsRwta93TbwNYoGKsUSChN44lgBNCoHLHzquYKrU6qZ8lolCIN0Rh6cP0Q3U6I6IXILYOQI513hJaSKAorFpuHXJNfVlpRtmYBk1Su1obZr5dnKAO+L10Hrj3WZW+E3qh6IszE37F6EB+68mGpvKm4eb9bFrlzrok7fvr0Kfv727dvWRmdVTJHw0qiiCUSZ6wCK+7XL/AcsgNyL74DQQ730sv78Su7+t/A36MdY0sW5o40ahslXr58aZ5HtZB8GH64m9EmMZ7FpYw4T6QnrZfgenrhFxaSiSGXtPnz57e9TkNZLvTjeqhr734CNtrK41L40sUQckmj1lGKQ0rC37x544r8eNXRpnVE3ZZY7zXo8NomiO0ZUCj2uHz58rbXoZ6gc0uA+F6ZeKS/jhRDUq8MKrTho9fEkihMmhxtBI1DxKFY9XLpVcSkfoi8JGnToZO5sU5aiDQIW716ddt7ZLYtMQlhECdBGXZZMWldY5BHm5xgAroWj4C0hbYkSc/jBmggIrXJWlZM6pSETsEPGqZOndr2uuuR5rF169a2HoHPdurUKZM4CO1WTPqaDaAd+GFGKdIQkxAn9RuEWcTRyN2KSUgiSgF5aWzPTeA/lN5rZubMmR2bE4SIC4nJoltgAV/dVefZm72AtctUCJU2CMJ327hxY9t7EHbkyJFseq+EJSY16RPo3Dkq1kkr7+q0bNmyDuLQcZBEPYmHVdOBiJyIlrRDq41YPWfXOxUysi5fvtyaj+2BpcnsUV/oSoEMOk2CQGlr4ckhBwaetBhjCwH0ZHtJROPJkyc7UjcYLDjmrH7ADTEBXFfOYmB0k9oYBOjJ8b4aOYSe7QkKcYhFlq3QYLQhSidNmtS2RATwy8YOM3EQJsUjKiaWZ+vZToUQgzhkHXudb/PW5YMHD9yZM2faPsMwoc7RciYJXbGuBqJ1UIGKKLv915jsvgtJxCZDubdXr165mzdvtr1Hz5LONA8jrUwKPqsmVesKa49S3Q4WxmRPUEYdTjgiUcfUwLx589ySJUva3oMkP6IYddq6HMS4o55xBJ
							 
						 
					
						
							
								
									
										
										
										
											2024-08-03 04:05:57 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								            }
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								          }
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        ]
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								      }
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    ],
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    "max_tokens": 300
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								  }'
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
									
										
										
										
											2024-08-03 04:16:23 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								curl http://localhost:11434/v1/completions \
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    -H "Content-Type: application/json" \
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    -d '{
							 
						 
					
						
							
								
									
										
										
										
											2024-09-26 02:11:22 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								        "model": "llama3.2",
							 
						 
					
						
							
								
									
										
										
										
											2024-08-03 04:16:23 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								        "prompt": "Say this is a test"
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    }'
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
									
										
										
										
											2024-08-02 06:48:44 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								curl http://localhost:11434/v1/models
							 
						 
					
						
							
								
									
										
										
										
											2024-08-02 06:58:13 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
									
										
										
										
											2024-09-26 02:11:22 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								curl http://localhost:11434/v1/models/llama3.2
							 
						 
					
						
							
								
									
										
										
										
											2024-08-02 07:00:29 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								curl http://localhost:11434/v1/embeddings \
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    -H "Content-Type: application/json" \
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    -d '{
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        "model": "all-minilm",
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        "input": ["why is the sky blue?", "why is the grass green?"]
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    }'
							 
						 
					
						
							
								
									
										
										
										
											2024-02-08 06:24:29 +08:00 
										
									 
								 
							 
							
								
							 
							
								 
							 
							
							
								```
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								## Endpoints
 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								### `/v1/chat/completions`
 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								#### Supported features
 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  Chat completions
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  Streaming
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  JSON mode
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  Reproducible outputs
							 
						 
					
						
							
								
									
										
										
										
											2024-08-03 04:05:57 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								- [x]  Vision
							 
						 
					
						
							
								
									
										
										
										
											2024-12-08 14:16:21 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								- [x]  Tools
							 
						 
					
						
							
								
									
										
										
										
											2024-02-08 06:24:29 +08:00 
										
									 
								 
							 
							
								
							 
							
								 
							 
							
							
								- [ ]  Logprobs
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								#### Supported request fields
 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  `model` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  `messages` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								  - [x]  Text `content` 
							 
						 
					
						
							
								
									
										
										
										
											2024-08-03 04:05:57 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								  - [x]  Image `content` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    - [x]  Base64 encoded image
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    - [ ]  Image URL
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								  - [x]  Array of `content`  parts
							 
						 
					
						
							
								
									
										
										
										
											2024-02-08 06:24:29 +08:00 
										
									 
								 
							 
							
								
							 
							
								 
							 
							
							
								- [x]  `frequency_penalty` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  `presence_penalty` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  `response_format` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  `seed` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  `stop` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  `stream` 
							 
						 
					
						
							
								
									
										
										
										
											2024-12-13 09:09:30 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								- [x]  `stream_options` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								  - [x]  `include_usage` 
							 
						 
					
						
							
								
									
										
										
										
											2024-02-08 06:24:29 +08:00 
										
									 
								 
							 
							
								
							 
							
								 
							 
							
							
								- [x]  `temperature` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  `top_p` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  `max_tokens` 
							 
						 
					
						
							
								
									
										
										
										
											2024-07-26 06:34:06 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								- [x]  `tools` 
							 
						 
					
						
							
								
									
										
										
										
											2024-02-08 06:24:29 +08:00 
										
									 
								 
							 
							
								
							 
							
								 
							 
							
							
								- [ ]  `tool_choice` 
							 
						 
					
						
							
								
									
										
										
										
											2024-07-26 06:34:06 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								- [ ]  `logit_bias` 
							 
						 
					
						
							
								
									
										
										
										
											2024-02-08 06:24:29 +08:00 
										
									 
								 
							 
							
								
							 
							
								 
							 
							
							
								- [ ]  `user` 
							 
						 
					
						
							
								
									
										
										
										
											2024-02-09 04:03:23 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								- [ ]  `n` 
							 
						 
					
						
							
								
									
										
										
										
											2024-02-08 06:24:29 +08:00 
										
									 
								 
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
									
										
										
										
											2024-08-03 04:16:23 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								### `/v1/completions`
 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								#### Supported features
 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  Completions
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  Streaming
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  JSON mode
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  Reproducible outputs
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [ ]  Logprobs
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								#### Supported request fields
 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  `model` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  `prompt` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  `frequency_penalty` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  `presence_penalty` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  `seed` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  `stop` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  `stream` 
							 
						 
					
						
							
								
									
										
										
										
											2024-12-13 09:09:30 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								- [x]  `stream_options` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								  - [x]  `include_usage` 
							 
						 
					
						
							
								
									
										
										
										
											2024-08-03 04:16:23 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								- [x]  `temperature` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  `top_p` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  `max_tokens` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  `suffix` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [ ]  `best_of` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [ ]  `echo` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [ ]  `logit_bias` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [ ]  `user` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [ ]  `n` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								#### Notes
 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								-  `prompt`  currently only accepts a string
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
									
										
										
										
											2024-08-02 06:48:44 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								### `/v1/models`
 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								#### Notes
 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								-  `created`  corresponds to when the model was last modified
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								-  `owned_by`  corresponds to the ollama username, defaulting to `"library"` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
									
										
										
										
											2024-08-02 06:58:13 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								### `/v1/models/{model}`
 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								#### Notes
 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								-  `created`  corresponds to when the model was last modified
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								-  `owned_by`  corresponds to the ollama username, defaulting to `"library"` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
									
										
										
										
											2024-08-02 07:00:29 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								### `/v1/embeddings`
 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								#### Supported request fields
 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  `model` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [x]  `input` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								  - [x]  string
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								  - [x]  array of strings
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								  - [ ]  array of tokens
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								  - [ ]  array of token arrays
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [ ]  `encoding format` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [ ]  `dimensions` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								- [ ]  `user` 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
									
										
										
										
											2024-02-08 06:24:29 +08:00 
										
									 
								 
							 
							
								
							 
							
								 
							 
							
							
								## Models
 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								Before using a model, pull it locally `ollama pull` :
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								```shell
							 
						 
					
						
							
								
									
										
										
										
											2024-09-26 02:11:22 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								ollama pull llama3.2
							 
						 
					
						
							
								
									
										
										
										
											2024-02-08 06:24:29 +08:00 
										
									 
								 
							 
							
								
							 
							
								 
							 
							
							
								```
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								### Default model names
 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								For tooling that relies on default OpenAI model names such as `gpt-3.5-turbo` , use `ollama cp`  to copy an existing model name to a temporary name:
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
									
										
										
										
											2025-02-08 01:55:07 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								```shell
							 
						 
					
						
							
								
									
										
										
										
											2024-09-26 02:11:22 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								ollama cp llama3.2 gpt-3.5-turbo
							 
						 
					
						
							
								
									
										
										
										
											2024-02-08 06:24:29 +08:00 
										
									 
								 
							 
							
								
							 
							
								 
							 
							
							
								```
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								Afterwards, this new model name can be specified the `model`  field:
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								```shell
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								curl http://localhost:11434/v1/chat/completions \
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    -H "Content-Type: application/json" \
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    -d '{
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        "model": "gpt-3.5-turbo",
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        "messages": [
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								            {
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								                "role": "user",
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								                "content": "Hello!"
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								            }
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        ]
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    }'
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								```
							 
						 
					
						
							
								
									
										
										
										
											2024-08-29 08:11:46 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								### Setting the context size
 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								The OpenAI API does not have a way of setting the context size for a model. If you need to change the context size, create a `Modelfile`  which looks like:
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
									
										
										
										
											2025-02-08 01:55:07 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								```
							 
						 
					
						
							
								
									
										
										
										
											2024-08-29 08:11:46 +08:00 
										
									 
								 
							 
							
								
									
										 
									 
								
							 
							
								 
							 
							
							
								FROM < some  model > 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								PARAMETER num_ctx < context  size > 
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								```
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								Use the `ollama create mymodel`  command to create a new model with the updated context size. Call the API with the updated model name:
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								```shell
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								curl http://localhost:11434/v1/chat/completions \
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    -H "Content-Type: application/json" \
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    -d '{
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        "model": "mymodel",
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        "messages": [
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								            {
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								                "role": "user",
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								                "content": "Hello!"
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								            }
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								        ]
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								    }'
							 
						 
					
						
							
								
							 
							
								
							 
							
								 
							 
							
							
								```