-
Notifications
You must be signed in to change notification settings - Fork 8.6k
Insights: ollama/ollama
Overview
Could not load contribution data
Please try again later
4 Pull requests merged by 2 people
-
make the modelfile path relative for
ollama create
#8380 merged
Jan 11, 2025 -
chore: upgrade to gods v2
#8196 merged
Jan 10, 2025 -
Mxyng/next llama
#8194 merged
Jan 10, 2025 -
merge llama/ggml into ml/backend/ggml
#7957 merged
Jan 10, 2025
3 Pull requests opened by 3 people
-
Explicit mention `ollama serve` will start a server, friendly for new users
#8381 opened
Jan 10, 2025 -
fix: add rootless container support
#8383 opened
Jan 11, 2025 -
add new create api doc
#8388 opened
Jan 12, 2025
47 Issues closed by 10 people
-
Run phi4 failed
#8370 closed
Jan 13, 2025 -
Llava embeddings: string into Go value of type api.EmbeddingRequest
#5861 closed
Jan 13, 2025 -
Error: max retries exceeded for all ollama model pulls (read: connection reset by peer)
#8167 closed
Jan 13, 2025 -
More API compatibility
#8251 closed
Jan 13, 2025 -
Ollama v0.5.4 not response with stream mode when submit tool option
#8278 closed
Jan 13, 2025 -
Ollama cannot load model after several hours on some GPUs
#8276 closed
Jan 13, 2025 -
Ollama models give low inference with Continue extension on VS Code Community Edition.
#8272 closed
Jan 13, 2025 -
Enhanced System Observability for Multi-Server Environments (Unified Endpoints?)
#8247 closed
Jan 13, 2025 -
Significant Performance Differences of ARM64 Installation Package Across Different CPU Architectures
#8246 closed
Jan 13, 2025 -
Error: open config.json: file does not exist
#8245 closed
Jan 13, 2025 -
ollama._types.ResponseError: llama runner process has terminated: signal: broken pipe
#8216 closed
Jan 13, 2025 -
Ollama hangs when running llama3.2 and llama3.2:1b
#8200 closed
Jan 13, 2025 -
Check Available Memory Before Downloading
#8192 closed
Jan 13, 2025 -
mllama doesn't support parallel requests yet - llama3.2-vision:11b for Standard_NC24ads_A100_v4
#8186 closed
Jan 13, 2025 -
Falcon3 10B in 1.58bit format
#8184 closed
Jan 13, 2025 -
How do I specify specific GPUs when running a model?
#8183 closed
Jan 13, 2025 -
{"error":"POST predict: Post \"http://127.0.0.1:33603/completion\": EOF"}
#8182 closed
Jan 13, 2025 -
LLAMA 3:70B is crashing inside K8s pods
#8179 closed
Jan 13, 2025 -
c4ai-command-r7b-12-2024
#8104 closed
Jan 13, 2025 -
ollama connection error on windows
#8075 closed
Jan 13, 2025 -
Ollama behind proxy can't pull new models anymore
#8050 closed
Jan 13, 2025 -
Ollama run very very slow in ARM cpu (KunPeng 920 CPU)
#8025 closed
Jan 13, 2025 -
Ollama is very slow after running for a while
#8023 closed
Jan 13, 2025 -
Web browsing
#8015 closed
Jan 13, 2025 -
Performance decline
#7919 closed
Jan 13, 2025 -
Installing bolt.new and qwen2.5-coder:7b locally (error cudaMalloc failed: out of memory)
#7896 closed
Jan 13, 2025 -
Proper way to train model on my data and load into Ollama?
#7755 closed
Jan 13, 2025 -
Nvidia fallback memory
#7584 closed
Jan 13, 2025 -
Despite advertised, granite3-dense does not seem to support tools.
#7544 closed
Jan 13, 2025 -
mac Errors when running
#7495 closed
Jan 13, 2025 -
Add Tab-Enabled Autocomplete for Local Model Parameters in Ollama CLI
#7239 closed
Jan 13, 2025 -
OLMo-2-1124-13B & 7B
#7863 closed
Jan 12, 2025 -
Ollama not completing chat request
#8387 closed
Jan 12, 2025 -
API is not giving up memory after responce is complete
#8390 closed
Jan 12, 2025 -
Hangs after 20-30 mins, a perdiocal restart of the ollama service is required
#6380 closed
Jan 12, 2025 -
How can we make model calls faster
#4493 closed
Jan 12, 2025 -
Setting up HTTP Server Timeouts / Connection Management
#3225 closed
Jan 12, 2025 -
Connection Error with OllamaFunctions in Langchain
#2783 closed
Jan 12, 2025 -
Error: llama runner process has terminated: exit status 2
#8382 closed
Jan 11, 2025 -
Unable to run llama on IPv6 Single Stack env
#8346 closed
Jan 11, 2025 -
GPU runs at maximum load with 2 models
#8285 closed
Jan 11, 2025 -
FROM path resolution uses working directory instead of Modelfile location
#8353 closed
Jan 11, 2025 -
Allow context to be set from the command line.
#8356 closed
Jan 11, 2025 -
Faiss document dependency error
#8357 closed
Jan 11, 2025 -
Add tools support to dolphin3
#8364 closed
Jan 10, 2025 -
deepseek v3
#8366 closed
Jan 10, 2025
18 Issues opened by 18 people
-
Sky-T1-32B- Preview would be a great model to add
#8403 opened
Jan 13, 2025 -
Allow building with BLAS/BLIS now that Ollama's runners are not pure native builds of llama.cpp anymore
#8402 opened
Jan 13, 2025 -
Failed to summarize the long context
#8401 opened
Jan 13, 2025 -
Model pulling behind proxy index out of range
#8400 opened
Jan 13, 2025 -
unable to use nvidia GPU & how to fix
#8399 opened
Jan 13, 2025 -
Cancelling model loading for one requests cancels it for others
#8398 opened
Jan 13, 2025 -
[UNK_BYTE_…] Output with gemma-2b-it in Ollama
#8397 opened
Jan 12, 2025 -
Error: could not connect to ollama app, is it running?
#8396 opened
Jan 12, 2025 -
Empty response via API
#8395 opened
Jan 12, 2025 -
Unable to enable GPU for models
#8393 opened
Jan 12, 2025 -
Empty 'assistant' message
#8392 opened
Jan 12, 2025 -
Add MoonDream 2 rev:2025-1-9 support
#8391 opened
Jan 12, 2025 -
Ollama install script relaces the systemd profile
#8389 opened
Jan 12, 2025 -
Return in a response a flag if the input request was truncated
#8386 opened
Jan 11, 2025 -
Cannot list or install models: Connection refused error on Windows 10
#8385 opened
Jan 11, 2025 -
Unable to acess ollama model hosted on a raspberry pi 5 from an other device
#8384 opened
Jan 11, 2025 -
Inquiry About Adding vikhyatk/moondream2 to Ollama
#8379 opened
Jan 10, 2025
26 Unresolved conversations
Sometimes conversations happen on old items that aren’t yet closed. Here is a list of all the Issues and Pull Requests with unresolved conversations.
-
examples: remove codified examples
#8267 commented on
Jan 13, 2025 • 2 new comments -
Remove redundant image_url format check in fromChatRequest
#8232 commented on
Jan 13, 2025 • 1 new comment -
Runner for OIlama engine
#8301 commented on
Jan 11, 2025 • 0 new comments -
create a default, non-root user for the container image
#8259 commented on
Jan 13, 2025 • 0 new comments -
fix: only add to history if different
#8197 commented on
Jan 10, 2025 • 0 new comments -
Examples(privategpt): update outdated langchain dependencies and add UnstructuredRSTLoader
#8117 commented on
Jan 13, 2025 • 0 new comments -
Support Unix sockets for OLLAMA_HOST
#8072 commented on
Jan 12, 2025 • 0 new comments -
wip: next ollama runner build updates
#7954 commented on
Jan 11, 2025 • 0 new comments -
wip: next ollama runner
#7913 commented on
Jan 11, 2025 • 0 new comments -
Add Vulkan support to ollama
#5059 commented on
Jan 12, 2025 • 0 new comments -
Deepseek (various) 236b crashes on run
#7867 commented on
Jan 13, 2025 • 0 new comments -
DeepSeek v3
#8249 commented on
Jan 13, 2025 • 0 new comments -
error looking up nvidia GPU memory - intermittent "cuda driver library failed to get device context 800"
#6928 commented on
Jan 13, 2025 • 0 new comments -
Corrupt output on multiple GPU in Windows 11
#8188 commented on
Jan 13, 2025 • 0 new comments -
Please don’t clutter the user home directory
#228 commented on
Jan 13, 2025 • 0 new comments -
Error: could not connect to ollama app, is it running?
#7524 commented on
Jan 13, 2025 • 0 new comments -
Rate limit download speed on pulling new models
#2006 commented on
Jan 12, 2025 • 0 new comments -
Download slows to a crawl at 99%
#1736 commented on
Jan 12, 2025 • 0 new comments -
Support model allenai/OLMo-7B
#2337 commented on
Jan 12, 2025 • 0 new comments -
Does ollama support accelerated running on npu?
#3004 commented on
Jan 12, 2025 • 0 new comments -
Support partial loads of LLaMA 3.2 Vision 11b on 6G GPUs
#7509 commented on
Jan 12, 2025 • 0 new comments -
llama3.2-vision doesn't utilize my GPU.
#8310 commented on
Jan 12, 2025 • 0 new comments -
When I use multiple GPUs, the utilization is very low.How can I configure it to maximize GPU utilization and reduce the reasoning time?
#8365 commented on
Jan 11, 2025 • 0 new comments -
Qwen/Qwen2.5-Math
#6889 commented on
Jan 11, 2025 • 0 new comments -
ollama not working
#8371 commented on
Jan 10, 2025 • 0 new comments -
Ollama version doesn't properly truncate tokens to 512 max for official snowflake-arctic-embed-l model
#8376 commented on
Jan 10, 2025 • 0 new comments