Skip to content

Commit ca5fbbd

Browse files
committed
chore: run go fmt
Signed-off-by: Jacob Howard <jacob.howard@docker.com>
1 parent ed476dc commit ca5fbbd

3 files changed

Lines changed: 4 additions & 4 deletions

File tree

pkg/inference/backends/llamacpp/download.go

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -25,9 +25,9 @@ const (
2525
)
2626

2727
var (
28-
ShouldUseGPUVariant bool
28+
ShouldUseGPUVariant bool
2929
ShouldUseGPUVariantLock sync.Mutex
30-
errLlamaCppUpToDate = errors.New("bundled llama.cpp version is up to date, no need to update")
30+
errLlamaCppUpToDate = errors.New("bundled llama.cpp version is up to date, no need to update")
3131
)
3232

3333
func (l *llamaCpp) downloadLatestLlamaCpp(ctx context.Context, log logging.Logger, httpClient *http.Client,

pkg/inference/backends/llamacpp/gpuinfo_notwindows.go

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,4 +4,4 @@ package llamacpp
44

55
import "context"
66

7-
func CanUseGPU(context.Context, string) (bool, error) { return false, nil }
7+
func CanUseGPU(context.Context, string) (bool, error) { return false, nil }

pkg/inference/backends/mlx/mlx.go

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -57,4 +57,4 @@ func (m *mlx) Run(ctx context.Context, socket, model string, mode inference.Back
5757

5858
func (m *mlx) Status() string {
5959
return "not running"
60-
}
60+
}

0 commit comments

Comments
 (0)