Skip to content

Commit 69102a2

Browse files
authored
small edits and placeholder gif (#2513)
* small edits and placeholder gif Signed-off-by: Max Cembalest <max@nomic.ai> * jul2 docs updates Signed-off-by: Max Cembalest <max@nomic.ai> * added video Signed-off-by: mcembalest <70534565+mcembalest@users.noreply.github.com> Signed-off-by: Max Cembalest <max@nomic.ai> * quantization nits Signed-off-by: Max Cembalest <max@nomic.ai> --------- Signed-off-by: Max Cembalest <max@nomic.ai> Signed-off-by: mcembalest <70534565+mcembalest@users.noreply.github.com>
1 parent b7d1b93 commit 69102a2

File tree

7 files changed

+61
-60
lines changed

7 files changed

+61
-60
lines changed

README.md

Lines changed: 14 additions & 31 deletions
Original file line numberDiff line numberDiff line change
@@ -2,6 +2,7 @@
22

33
<p align="center">GPT4All runs large language models (LLMs) privately on everyday desktops & laptops. <br> <br> No API calls or GPUs required - you can just download the application and <a href="https://docs.gpt4all.io/gpt4all_desktop/quickstart.html#quickstart">get started</a>
44

5+
https://github.com/nomic-ai/gpt4all/assets/70534565/513a0f15-4964-4109-89e4-4f9a9011f311
56

67
<p align="center">
78
<a href="https://gpt4all.io/installers/gpt4all-installer-win64.exe">
@@ -12,15 +13,15 @@
1213

1314
<p align="center">
1415
<a href="https://gpt4all.io/installers/gpt4all-installer-darwin.dmg">
15-
<img src="gpt4all-bindings/python/docs/assets/mac.png" width="80" height="90"><br>
16+
<img src="gpt4all-bindings/python/docs/assets/mac.png" width="85" height="100"><br>
1617
Download for MacOS
1718
</a>
1819
</p>
1920

2021
<p align="center">
2122
<a href="https://gpt4all.io/installers/gpt4all-installer-linux.run">
22-
<img src="gpt4all-bindings/python/docs/assets/linux.png" width="80" height="80"><br>
23-
Download for Linux
23+
<img src="gpt4all-bindings/python/docs/assets/ubuntu.svg" width="120" height="120"><br>
24+
Download for Ubuntu
2425
</a>
2526
</p>
2627

@@ -37,8 +38,6 @@ GPT4All is made possible by our compute partner <a href="https://www.paperspace.
3738
<a href="https://www.phorm.ai/query?projectId=755eecd3-24ad-49cc-abf4-0ab84caacf63"><img src="https://img.shields.io/badge/Phorm-Ask_AI-%23F2777A.svg" alt="phorm.ai"></a>
3839
</p>
3940

40-
41-
4241
## Install GPT4All Python
4342

4443
`gpt4all` gives you access to LLMs with our Python client around [`llama.cpp`](https://github.com/ggerganov/llama.cpp) implementations.
@@ -57,10 +56,17 @@ with model.chat_session():
5756
```
5857

5958

60-
### Release History
59+
## Integrations
60+
61+
:parrot::link: [Langchain](https://python.langchain.com/v0.2/docs/integrations/providers/gpt4all/)
62+
:card_file_box: [Weaviate Vector Database](https://github.com/weaviate/weaviate) - [module docs](https://weaviate.io/developers/weaviate/modules/retriever-vectorizer-modules/text2vec-gpt4all)
63+
:telescope: [OpenLIT (OTel-native Monitoring)](https://github.com/openlit/openlit) - [Docs](https://docs.openlit.io/latest/integrations/gpt4all)
64+
65+
## Release History
6166
- **July 2nd, 2024**: V3.0.0 Release
62-
- New UI/UX: fresh redesign of the chat application GUI and user experience
63-
- LocalDocs: bring information from files on-device into chats
67+
- Fresh redesign of the chat application UI
68+
- Improved user workflow for LocalDocs
69+
- Expanded access to more model architectures
6470
- **October 19th, 2023**: GGUF Support Launches with Support for:
6571
- Mistral 7b base model, an updated model gallery on [gpt4all.io](https://gpt4all.io), several new local code models including Rift Coder v1.5
6672
- [Nomic Vulkan](https://blog.nomic.ai/posts/gpt4all-gpu-inference-with-vulkan) support for Q4\_0 and Q4\_1 quantizations in GGUF.
@@ -71,13 +77,6 @@ with model.chat_session():
7177

7278
[Docker-based API server]: https://github.com/nomic-ai/gpt4all/tree/cef74c2be20f5b697055d5b8b506861c7b997fab/gpt4all-api
7379

74-
### Integrations
75-
76-
* :parrot::link: [Langchain](https://python.langchain.com/v0.2/docs/integrations/providers/gpt4all/)
77-
* :card_file_box: [Weaviate Vector Database](https://github.com/weaviate/weaviate) - [module docs](https://weaviate.io/developers/weaviate/modules/retriever-vectorizer-modules/text2vec-gpt4all)
78-
* :telescope: [OpenLIT (OTel-native Monitoring)](https://github.com/openlit/openlit) - [Docs](https://docs.openlit.io/latest/integrations/gpt4all)
79-
80-
8180
## Contributing
8281
GPT4All welcomes contributions, involvement, and discussion from the open source community!
8382
Please see CONTRIBUTING.md and follow the issues, bug reports, and PR markdown templates.
@@ -86,22 +85,6 @@ Check project discord, with project owners, or through existing issues/PRs to av
8685
Please make sure to tag all of the above with relevant project identifiers or your contribution could potentially get lost.
8786
Example tags: `backend`, `bindings`, `python-bindings`, `documentation`, etc.
8887

89-
90-
## Technical Reports
91-
92-
<p align="center">
93-
<a href="https://gpt4all.io/reports/GPT4All_Technical_Report_3.pdf">:green_book: Technical Report 3: GPT4All Snoozy and Groovy </a>
94-
</p>
95-
96-
<p align="center">
97-
<a href="https://static.nomic.ai/gpt4all/2023_GPT4All-J_Technical_Report_2.pdf">:green_book: Technical Report 2: GPT4All-J </a>
98-
</p>
99-
100-
<p align="center">
101-
<a href="https://s3.amazonaws.com/static.nomic.ai/gpt4all/2023_GPT4All_Technical_Report.pdf">:green_book: Technical Report 1: GPT4All</a>
102-
</p>
103-
104-
10588
## Citation
10689

10790
If you utilize this repository, models or data in a downstream project, please consider citing it with:
Lines changed: 5 additions & 0 deletions
Loading

gpt4all-bindings/python/docs/gpt4all_desktop/models.md

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -56,13 +56,13 @@ Many LLMs are available at various sizes, quantizations, and licenses.
5656

5757
Here are a few examples:
5858

59-
| Model| Filesize| RAM Required| Parameters| Developer| License| MD5 Sum (Unique Hash)|
60-
|------|---------|-------------|-----------|----------|--------|----------------------|
61-
| Llama 3 Instruct | 4.66 GB| 8 GB| 8 Billion| Meta| [Llama 3 License](https://llama.meta.com/llama3/license/)| c87ad09e1e4c8f9c35a5fcef52b6f1c9|
62-
| Nous Hermes 2 Mistral DPO| 4.21 GB| 8 GB| 7 Billion| Mistral & Nous Research | [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0)| Coa5f6b4eabd3992da4d7fb7f020f921eb|
63-
| Phi-3 Mini Instruct | 2.03 GB| 4 GB| 4 billion| Microsoft| [MIT](https://opensource.org/license/mit)| f8347badde9bfc2efbe89124d78ddaf5|
64-
| Mini Orca (Small)| 1.84 GB| 4 GB| 3 billion| Microsoft | [CC-BY-NC-SA-4.0](https://spdx.org/licenses/CC-BY-NC-SA-4.0)| 0e769317b90ac30d6e09486d61fefa26|
65-
| GPT4All Snoozy| 7.36 GB| 16 GB| 13 billion| Nomic AI| [GPL](https://www.gnu.org/licenses/gpl-3.0.en.html)| 40388eb2f8d16bb5d08c96fdfaac6b2c|
59+
| Model| Filesize| RAM Required| Parameters| Quantization| Developer| License| MD5 Sum (Unique Hash)|
60+
|------|---------|-------------|-----------|-------------|----------|--------|----------------------|
61+
| Llama 3 Instruct | 4.66 GB| 8 GB| 8 Billion| q4_0| Meta| [Llama 3 License](https://llama.meta.com/llama3/license/)| c87ad09e1e4c8f9c35a5fcef52b6f1c9|
62+
| Nous Hermes 2 Mistral DPO| 4.11 GB| 8 GB| 7 Billion| q4_0| Mistral & Nous Research | [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0)| Coa5f6b4eabd3992da4d7fb7f020f921eb|
63+
| Phi-3 Mini Instruct | 2.18 GB| 4 GB| 4 billion| q4_0| Microsoft| [MIT](https://opensource.org/license/mit)| f8347badde9bfc2efbe89124d78ddaf5|
64+
| Mini Orca (Small)| 1.98 GB| 4 GB| 3 billion| q4_0| Microsoft | [CC-BY-NC-SA-4.0](https://spdx.org/licenses/CC-BY-NC-SA-4.0)| 0e769317b90ac30d6e09486d61fefa26|
65+
| GPT4All Snoozy| 7.37 GB| 16 GB| 13 billion| q4_0| Nomic AI| [GPL](https://www.gnu.org/licenses/gpl-3.0.en.html)| 40388eb2f8d16bb5d08c96fdfaac6b2c|
6666

6767
### Search Results
6868

gpt4all-bindings/python/docs/gpt4all_help/faq.md

Lines changed: 2 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -4,17 +4,11 @@
44

55
### Which language models are supported?
66

7-
Our backend supports models with a `llama.cpp` implementation which have been uploaded to [HuggingFace](https://huggingface.co/).
7+
We support models with a `llama.cpp` implementation which have been uploaded to [HuggingFace](https://huggingface.co/).
88

99
### Which embedding models are supported?
1010

11-
The following embedding models can be used within the application and with the `Embed4All` class from the `gpt4all` Python library. The default context length as GGUF files is 2048 but can be [extended](https://huggingface.co/nomic-ai/nomic-embed-text-v1.5-GGUF#description).
12-
13-
| Name | Initializing with `Embed4All` | Context Length | Embedding Length | File Size |
14-
|--------------------|------------------------------------------------------|---------------:|-----------------:|----------:|
15-
| [SBert](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2)| ```pythonemb = Embed4All("all-MiniLM-L6-v2.gguf2.f16.gguf")```| 512 | 384 | 44 MiB |
16-
| [Nomic Embed v1](https://huggingface.co/nomic-ai/nomic-embed-text-v1-GGUF) | nomic&#x2011;embed&#x2011;text&#x2011;v1.f16.gguf| 2048 | 768 | 262 MiB |
17-
| [Nomic Embed v1.5](https://huggingface.co/nomic-ai/nomic-embed-text-v1.5-GGUF) | nomic&#x2011;embed&#x2011;text&#x2011;v1.5.f16.gguf| 2048 | 64-768 | 262 MiB |
11+
We support SBert and Nomic Embed Text v1 & v1.5.
1812

1913
## Software
2014

gpt4all-bindings/python/docs/gpt4all_python/home.md

Lines changed: 17 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -23,6 +23,15 @@ Models are loaded by name via the `GPT4All` class. If it's your first time loadi
2323
print(model.generate("How can I run LLMs efficiently on my laptop?", max_tokens=1024))
2424
```
2525

26+
| `GPT4All` model name| Filesize| RAM Required| Parameters| Quantization| Developer| License| MD5 Sum (Unique Hash)|
27+
|------|---------|-------|-------|-----------|----------|--------|----------------------|
28+
| `Meta-Llama-3-8B-Instruct.Q4_0.gguf`| 4.66 GB| 8 GB| 8 Billion| q4_0| Meta| [Llama 3 License](https://llama.meta.com/llama3/license/)| c87ad09e1e4c8f9c35a5fcef52b6f1c9|
29+
| `Nous-Hermes-2-Mistral-7B-DPO.Q4_0.gguf`| 4.11 GB| 8 GB| 7 Billion| q4_0| Mistral & Nous Research | [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0)| Coa5f6b4eabd3992da4d7fb7f020f921eb|
30+
| `Phi-3-mini-4k-instruct.Q4_0.gguf` | 2.18 GB| 4 GB| 3.8 billion| q4_0| Microsoft| [MIT](https://opensource.org/license/mit)| f8347badde9bfc2efbe89124d78ddaf5|
31+
| `orca-mini-3b-gguf2-q4_0.gguf`| 1.98 GB| 4 GB| 3 billion| q4_0| Microsoft | [CC-BY-NC-SA-4.0](https://spdx.org/licenses/CC-BY-NC-SA-4.0)| 0e769317b90ac30d6e09486d61fefa26|
32+
| `gpt4all-13b-snoozy-q4_0.gguf`| 7.37 GB| 16 GB| 13 billion| q4_0| Nomic AI| [GPL](https://www.gnu.org/licenses/gpl-3.0.en.html)| 40388eb2f8d16bb5d08c96fdfaac6b2c|
33+
34+
2635
## Chat Session Generation
2736

2837
Most of the language models you will be able to access from HuggingFace have been trained as assistants. This guides language models to not just answer with relevant text, but *helpful* text.
@@ -75,16 +84,6 @@ If you want your LLM's responses to be helpful in the typical sense, we recommen
7584
b = 5
7685
```
7786

78-
## Example Models
79-
80-
| Model| Filesize| RAM Required| Parameters| Developer| License| MD5 Sum (Unique Hash)|
81-
|------|---------|-------------|-----------|----------|--------|----------------------|
82-
| `Meta-Llama-3-8B-Instruct.Q4_0.gguf` | 4.66 GB| 8 GB| 8 Billion| Meta| [Llama 3 License](https://llama.meta.com/llama3/license/)| c87ad09e1e4c8f9c35a5fcef52b6f1c9|
83-
| Nous Hermes 2 Mistral DPO| 4.21 GB| 8 GB| 7 Billion| Mistral & Nous Research | [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0)| Coa5f6b4eabd3992da4d7fb7f020f921eb|
84-
| Phi-3 Mini Instruct | 2.03 GB| 4 GB| 4 billion| Microsoft| [MIT](https://opensource.org/license/mit)| f8347badde9bfc2efbe89124d78ddaf5|
85-
| Mini Orca (Small)| 1.84 GB| 4 GB| 3 billion| Microsoft | [CC-BY-NC-SA-4.0](https://spdx.org/licenses/CC-BY-NC-SA-4.0)| 0e769317b90ac30d6e09486d61fefa26|
86-
| GPT4All Snoozy| 7.36 GB| 16 GB| 13 billion| Nomic AI| [GPL](https://www.gnu.org/licenses/gpl-3.0.en.html)| 40388eb2f8d16bb5d08c96fdfaac6b2c|
87-
8887
## Direct Generation
8988

9089
Directly calling `model.generate()` prompts the model without applying any templates.
@@ -150,3 +149,11 @@ The easiest way to run the text embedding model locally uses the [`nomic`](https
150149
![Nomic embed text local inference](../assets/local_embed.gif)
151150

152151
To learn more about making embeddings locally with `nomic`, visit our [embeddings guide](https://docs.nomic.ai/atlas/guides/embeddings#local-inference).
152+
153+
The following embedding models can be used within the application and with the `Embed4All` class from the `gpt4all` Python library. The default context length as GGUF files is 2048 but can be [extended](https://huggingface.co/nomic-ai/nomic-embed-text-v1.5-GGUF#description).
154+
155+
| Name| Using with `nomic`| `Embed4All` model name| Context Length| # Embedding Dimensions| File Size|
156+
|--------------------|-|------------------------------------------------------|---------------:|-----------------:|----------:|
157+
| [Nomic Embed v1](https://huggingface.co/nomic-ai/nomic-embed-text-v1-GGUF) | ```embed.text(strings, model="nomic-embed-text-v1", inference_mode="local")```| ```Embed4All("nomic-embed-text-v1.f16.gguf")```| 2048 | 768 | 262 MiB |
158+
| [Nomic Embed v1.5](https://huggingface.co/nomic-ai/nomic-embed-text-v1.5-GGUF) | ```embed.text(strings, model="nomic-embed-text-v1.5", inference_mode="local")```| ```Embed4All("nomic-embed-text-v1.5.f16.gguf")``` | 2048| 64-768 | 262 MiB |
159+
| [SBert](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2)| n/a| ```Embed4All("all-MiniLM-L6-v2.gguf2.f16.gguf")```| 512 | 384 | 44 MiB |

gpt4all-training/README.md

Lines changed: 13 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,18 @@
11
## Training GPT4All-J
22

3-
Please see [GPT4All-J Technical Report](https://static.nomic.ai/gpt4all/2023_GPT4All-J_Technical_Report_2.pdf) for details.
3+
### Technical Reports
4+
5+
<p align="center">
6+
<a href="https://gpt4all.io/reports/GPT4All_Technical_Report_3.pdf">:green_book: Technical Report 3: GPT4All Snoozy and Groovy </a>
7+
</p>
8+
9+
<p align="center">
10+
<a href="https://static.nomic.ai/gpt4all/2023_GPT4All-J_Technical_Report_2.pdf">:green_book: Technical Report 2: GPT4All-J </a>
11+
</p>
12+
13+
<p align="center">
14+
<a href="https://s3.amazonaws.com/static.nomic.ai/gpt4all/2023_GPT4All_Technical_Report.pdf">:green_book: Technical Report 1: GPT4All</a>
15+
</p>
416

517
### GPT4All-J Training Data
618

roadmap.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -11,15 +11,15 @@ Each item should have an issue link below.
1111
- [ ] Portuguese
1212
- [ ] Your native language here.
1313
- UI Redesign: an internal effort at Nomic to improve the UI/UX of gpt4all for all users.
14-
- [ ] Design new user interface and gather community feedback
15-
- [ ] Implement the new user interface and experience.
14+
- [x] Design new user interface and gather community feedback
15+
- [x] Implement the new user interface and experience.
1616
- Installer and Update Improvements
1717
- [ ] Seamless native installation and update process on OSX
1818
- [ ] Seamless native installation and update process on Windows
1919
- [ ] Seamless native installation and update process on Linux
2020
- Model discoverability improvements:
2121
- [x] Support huggingface model discoverability
22-
- [ ] Support Nomic hosted model discoverability
22+
- [x] Support Nomic hosted model discoverability
2323
- LocalDocs (towards a local perplexity)
2424
- Multilingual LocalDocs Support
2525
- [ ] Create a multilingual experience

0 commit comments

Comments
 (0)