--- base_model: NewEden/DPE-70b-Ckpts datasets: - merged.json - PocketDoc/Dans-Prosemaxx-RP - PocketDoc/Dans-Personamaxx-Logs-2 - PocketDoc/Dans-Personamaxx-VN - PocketDoc/Dans-Kinomaxx-VanillaBackrooms - PocketDoc/Dans-Prosemaxx-Gutenberg - PocketDoc/Dans-Prosemaxx-Cowriter-3-XL - PocketDoc/Dans-Prosemaxx-Adventure - PocketDoc/Dans-Failuremaxx-Adventure-3 - PocketDoc/Dans-Prosemaxx-InstructWriter-ZeroShot-2 - PocketDoc/Dans-Prosemaxx-InstructWriter-ZeroShot-3 - PocketDoc/Dans-Prosemaxx-InstructWriter-Continue-2 - PocketDoc/Dans-Prosemaxx-Instructwriter-Long - PocketDoc/Dans-Prosemaxx-RepRemover-1 - PocketDoc/Dans-MemoryCore-CoreCurriculum-Small - AquaV/US-Army-Survival-Sharegpt - AquaV/Multi-Environment-Operations-Sharegpt - AquaV/Resistance-Sharegpt - AquaV/Interrogation-Sharegpt - AquaV/Chemical-Biological-Safety-Applications-Sharegpt - AquaV/Energetic-Materials-Sharegpt - PocketDoc/Dans-Mathmaxx - PJMixers/Math-Multiturn-1K-ShareGPT - PocketDoc/Dans-Taskmaxx - PocketDoc/Dans-Taskmaxx-DataPrepper - PocketDoc/Dans-Taskmaxx-ConcurrentQA-Reworked - PocketDoc/Dans-Taskmaxx-TableGPT - PocketDoc/Dans-Taskmaxx-SciRIFF - PocketDoc/Dans-Taskmaxx-Edit - PocketDoc/Dans-Toolmaxx-Agent - PocketDoc/Dans-Toolmaxx-ShellCommands - PocketDoc/Dans-Toolmaxx-Functions-Toolbench - PocketDoc/Dans-Toolmaxx-Functions-ToolACE - PocketDoc/Dans-Toolmaxx-Functions-apigen-subset - PocketDoc/Dans-Assistantmaxx-OpenAssistant2 - PocketDoc/Dans-Assistantmaxx-Opus-Merge-2 - PocketDoc/Dans-Assistantmaxx-sonnetorca-subset - PocketDoc/Dans-Assistantmaxx-sonnetorca-subset-2 - PocketDoc/Dans-Assistantmaxx-Synthia - PocketDoc/Dans-Assistantmaxx-ASL - PocketDoc/Dans-Assistantmaxx-PersonaLLM-Opus - PocketDoc/Dans-Assistantmaxx-LongAlign - PocketDoc/Dans-Assistantmaxx-OpenLeecher-Instruct - PocketDoc/Dans-Assistantmaxx-Tulu3-IF - PocketDoc/Dans-Systemmaxx - PocketDoc/Dans-Logicmaxx-SAT-AP - PJMixers/grimulkan_theory-of-mind-ShareGPT - PJMixers/grimulkan_physical-reasoning-ShareGPT - PocketDoc/Dans-Reasoningmaxx-NaturalReasoning - PocketDoc/Dans-Reasoningmaxx-WebInstruct - PocketDoc/Dans-Reasoningmaxx-GeneralReasoning - Delta-Vector/Orion-LN-V1-ShareGPT - Delta-Vector/Orion-Alpindale-LN-ShareGPT - Delta-Vector/Orion-Shoujo-AI-Filtered-ShareGPT - Delta-Vector/Orion-RP-Guild - Delta-Vector/Orion-OpenCAI-ShareGPT - Delta-Vector/Orion-LIMARP-Complexity language: - en library_name: transformers license: llama3.1 mradermacher: readme_rev: 1 quantized_by: mradermacher tags: - generated_from_trainer --- ## About static quants of https://huggingface.co/NewEden/DPE-70b-Ckpts ***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#DPE-70b-Ckpts-GGUF).*** weighted/imatrix quants seem not to be available (by me) at this time. If they do not show up a week or so after the static ones, I have probably not planned for them. Feel free to request them by opening a Community Discussion. ## Usage If you are unsure how to use GGUF files, refer to one of [TheBloke's READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for more details, including on how to concatenate multi-part files. ## Provided Quants (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | [GGUF](https://huggingface.co/mradermacher/DPE-70b-Ckpts-GGUF/resolve/main/DPE-70b-Ckpts.Q4_K_S.gguf) | Q4_K_S | 40.4 | fast, recommended | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): ![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 ## FAQ / Model Request See https://huggingface.co/mradermacher/model_requests for some answers to questions you might have and/or if you want some other model quantized. ## Thanks I thank my company, [nethype GmbH](https://www.nethype.de/), for letting me use its servers and providing upgrades to my workstation to enable this work in my free time.