Skip to content

Settings

Application Settings

General Application Settings

Setting Description Default Value
Theme Color theme for the application. Options are Light, Dark, and LegacyDark Light
Font Size Font size setting for text throughout the application. Options are Small, Medium, and Large Small
Device Device that will run your models. Options are Auto (GPT4All chooses), Metal (Apple Silicon M1+), CPU, and GPU Auto
Default Model Choose your preferred LLM to load by default on startup Auto
Download Path Select a destination on your device to save downloaded models Windows: C:\Users\{username}\AppData\Local\nomic.ai\GPT4All

Mac: /Users/{username}/Library/Application Support/nomic.ai/GPT4All/

Linux: /home/{username}/.local/share/nomic.ai/GPT4All
Enable Datalake Opt-in to sharing interactions with GPT4All community (anonymous and optional) Off

Advanced Application Settings

Setting Description Default Value
CPU Threads Number of concurrently running CPU threads (more can speed up responses) 4
Save Chat Context Save chat context to disk to pick up exactly where a model left off. Off
Enable Local Server Allow any application on your device to use GPT4All via an OpenAI-compatible GPT4All API Off
API Server Port Local HTTP port for the local API server 4891

Model Settings

Model / Character Settings

Setting Description Default Value
Name Unique name of this model / character set by model uploader
System Prompt General instructions for the chats this model will be used for set by model uploader
Prompt Template Format of user <-> assistant interactions for the chats this model will be used for set by model uploader

Clone

You can clone an existing model, which allows you to save a configuration of a model file with different prompt templates and sampling settings.

Sampling Settings

Model Sampling Settings

Setting Description Default Value
Context Length Maximum length of input sequence in tokens 2048
Max Length Maximum length of response in tokens 4096
Prompt Batch Size Token batch size for parallel processing 128
Temperature Lower temperature gives more likely generations 0.7
Top P Prevents choosing highly unlikely tokens 0.4
Top K Size of selection pool for tokens 40
Min P Minimum relative probability 0
Repeat Penalty Tokens Length to apply penalty 64
Repeat Penalty Penalize repetitiveness 1.18
GPU Layers How many model layers to load into VRAM 32

LocalDocs Settings

General LocalDocs Settings

Setting Description Default Value
Allowed File Extensions Choose which file types will be indexed into LocalDocs collections as text snippets with embedding vectors .txt, .pdf, .md, .rst
Use Nomic Embed API Use Nomic API to create LocalDocs collections fast and off-device; Nomic API Key required Off
Embeddings Device Device that will run embedding models. Options are Auto (GPT4All chooses), Metal (Apple Silicon M1+), CPU, and GPU Auto
Show Sources Titles of source files retrieved by LocalDocs will be displayed directly in your chats. On

Advanced LocalDocs Settings

Note that increasing these settings can increase the likelihood of factual responses, but may result in slower generation times.

Setting Description Default Value
Document Snippet Size Number of string characters per document snippet 512
Maximum Document Snippets Per Prompt Upper limit for the number of snippets from your files LocalDocs can retrieve for LLM context 3