mirror of https://github.com/ollama/ollama.git
				
				
				
			
		
			
				
	
	
		
			253 lines
		
	
	
		
			5.3 KiB
		
	
	
	
		
			Go
		
	
	
	
			
		
		
	
	
			253 lines
		
	
	
		
			5.3 KiB
		
	
	
	
		
			Go
		
	
	
	
| package server
 | |
| 
 | |
| import (
 | |
| 	"archive/zip"
 | |
| 	"bytes"
 | |
| 	"context"
 | |
| 	"errors"
 | |
| 	"fmt"
 | |
| 	"io"
 | |
| 	"net/http"
 | |
| 	"os"
 | |
| 	"path/filepath"
 | |
| 
 | |
| 	"github.com/ollama/ollama/api"
 | |
| 	"github.com/ollama/ollama/convert"
 | |
| 	"github.com/ollama/ollama/llm"
 | |
| 	"github.com/ollama/ollama/types/model"
 | |
| )
 | |
| 
 | |
| var intermediateBlobs map[string]string = make(map[string]string)
 | |
| 
 | |
| type layerWithGGML struct {
 | |
| 	*Layer
 | |
| 	*llm.GGML
 | |
| }
 | |
| 
 | |
| func parseFromModel(ctx context.Context, name model.Name, fn func(api.ProgressResponse)) (layers []*layerWithGGML, err error) {
 | |
| 	m, err := ParseNamedManifest(name)
 | |
| 	switch {
 | |
| 	case errors.Is(err, os.ErrNotExist):
 | |
| 		if err := PullModel(ctx, name.String(), ®istryOptions{}, fn); err != nil {
 | |
| 			return nil, err
 | |
| 		}
 | |
| 
 | |
| 		m, err = ParseNamedManifest(name)
 | |
| 		if err != nil {
 | |
| 			return nil, err
 | |
| 		}
 | |
| 	case err != nil:
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	for _, layer := range m.Layers {
 | |
| 		layer, err := NewLayerFromLayer(layer.Digest, layer.MediaType, name.DisplayShortest())
 | |
| 		if err != nil {
 | |
| 			return nil, err
 | |
| 		}
 | |
| 
 | |
| 		switch layer.MediaType {
 | |
| 		case "application/vnd.ollama.image.model",
 | |
| 			"application/vnd.ollama.image.projector",
 | |
| 			"application/vnd.ollama.image.adapter":
 | |
| 			blobpath, err := GetBlobsPath(layer.Digest)
 | |
| 			if err != nil {
 | |
| 				return nil, err
 | |
| 			}
 | |
| 
 | |
| 			blob, err := os.Open(blobpath)
 | |
| 			if err != nil {
 | |
| 				return nil, err
 | |
| 			}
 | |
| 			defer blob.Close()
 | |
| 
 | |
| 			ggml, _, err := llm.DecodeGGML(blob)
 | |
| 			if err != nil {
 | |
| 				return nil, err
 | |
| 			}
 | |
| 
 | |
| 			layers = append(layers, &layerWithGGML{layer, ggml})
 | |
| 		default:
 | |
| 			layers = append(layers, &layerWithGGML{layer, nil})
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return layers, nil
 | |
| }
 | |
| 
 | |
| func parseFromZipFile(_ context.Context, file *os.File, digest string, fn func(api.ProgressResponse)) (layers []*layerWithGGML, err error) {
 | |
| 	stat, err := file.Stat()
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	r, err := zip.NewReader(file, stat.Size())
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	tempdir, err := os.MkdirTemp(filepath.Dir(file.Name()), "")
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 	defer os.RemoveAll(tempdir)
 | |
| 
 | |
| 	fn(api.ProgressResponse{Status: "unpacking model metadata"})
 | |
| 	for _, f := range r.File {
 | |
| 		// TODO(mxyng): this should not write out all files to disk
 | |
| 		outfile, err := os.Create(filepath.Join(tempdir, f.Name))
 | |
| 		if err != nil {
 | |
| 			return nil, err
 | |
| 		}
 | |
| 		defer outfile.Close()
 | |
| 
 | |
| 		infile, err := f.Open()
 | |
| 		if err != nil {
 | |
| 			return nil, err
 | |
| 		}
 | |
| 		defer infile.Close()
 | |
| 
 | |
| 		if _, err = io.Copy(outfile, infile); err != nil {
 | |
| 			return nil, err
 | |
| 		}
 | |
| 
 | |
| 		if err := outfile.Close(); err != nil {
 | |
| 			return nil, err
 | |
| 		}
 | |
| 
 | |
| 		if err := infile.Close(); err != nil {
 | |
| 			return nil, err
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	mf, err := convert.GetModelFormat(tempdir)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	params, err := mf.GetParams(tempdir)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	mArch, err := mf.GetModelArch("", tempdir, params)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	fn(api.ProgressResponse{Status: "processing tensors"})
 | |
| 	if err := mArch.GetTensors(); err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	if err := mArch.LoadVocab(); err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	fn(api.ProgressResponse{Status: "converting model"})
 | |
| 
 | |
| 	// TODO(mxyng): this should write directly into a layer
 | |
| 	// e.g. NewLayer(arch.Reader(), "application/vnd.ollama.image.model")
 | |
| 	temp, err := os.CreateTemp(tempdir, "fp16")
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 	defer temp.Close()
 | |
| 	defer os.Remove(temp.Name())
 | |
| 
 | |
| 	if err = mArch.WriteGGUF(temp); err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	if _, err := temp.Seek(0, io.SeekStart); err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	layer, err := NewLayer(temp, "application/vnd.ollama.image.model")
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	bin, err := layer.Open()
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 	defer bin.Close()
 | |
| 
 | |
| 	ggml, _, err := llm.DecodeGGML(bin)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	layers = append(layers, &layerWithGGML{layer, ggml})
 | |
| 
 | |
| 	intermediateBlobs[digest] = layer.Digest
 | |
| 	return layers, nil
 | |
| }
 | |
| 
 | |
| func parseFromFile(ctx context.Context, file *os.File, digest string, fn func(api.ProgressResponse)) (layers []*layerWithGGML, err error) {
 | |
| 	sr := io.NewSectionReader(file, 0, 512)
 | |
| 	contentType, err := detectContentType(sr)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	switch contentType {
 | |
| 	case "gguf", "ggla":
 | |
| 		// noop
 | |
| 	case "application/zip":
 | |
| 		return parseFromZipFile(ctx, file, digest, fn)
 | |
| 	default:
 | |
| 		return nil, fmt.Errorf("unsupported content type: %s", contentType)
 | |
| 	}
 | |
| 
 | |
| 	stat, err := file.Stat()
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 
 | |
| 	var offset int64
 | |
| 	for offset < stat.Size() {
 | |
| 		ggml, n, err := llm.DecodeGGML(file)
 | |
| 		if errors.Is(err, io.EOF) {
 | |
| 			break
 | |
| 		} else if err != nil {
 | |
| 			return nil, err
 | |
| 		}
 | |
| 
 | |
| 		mediatype := "application/vnd.ollama.image.model"
 | |
| 		if ggml.Name() == "ggla" {
 | |
| 			mediatype = "application/vnd.ollama.image.adapter"
 | |
| 		} else if ggml.KV().Architecture() == "clip" {
 | |
| 			mediatype = "application/vnd.ollama.image.projector"
 | |
| 		}
 | |
| 
 | |
| 		layer, err := NewLayer(io.NewSectionReader(file, offset, n), mediatype)
 | |
| 		if err != nil {
 | |
| 			return nil, err
 | |
| 		}
 | |
| 
 | |
| 		layers = append(layers, &layerWithGGML{layer, ggml})
 | |
| 		offset = n
 | |
| 	}
 | |
| 
 | |
| 	return layers, nil
 | |
| }
 | |
| 
 | |
| func detectContentType(r io.Reader) (string, error) {
 | |
| 	var b bytes.Buffer
 | |
| 	if _, err := io.Copy(&b, r); err != nil {
 | |
| 		return "", err
 | |
| 	}
 | |
| 
 | |
| 	if contentType := llm.DetectGGMLType(b.Bytes()); contentType != "" {
 | |
| 		return contentType, nil
 | |
| 	}
 | |
| 
 | |
| 	if contentType := http.DetectContentType(b.Bytes()); contentType != "application/octet-stream" {
 | |
| 		return contentType, nil
 | |
| 	}
 | |
| 
 | |
| 	return "unknown", nil
 | |
| }
 |