Commit graph

  • bb7e410a0c Improve DRY sampler performance 2 of 2 (introduce Z algorithm). Belladore 2024-05-23 02:44:29 +0300
  • d9fdb3db71 Add docs for image generation Artificiangel 2024-05-23 08:44:15 -0400
  • ee7d2c7406 missing requirements for transcription endpoint Artificiangel 2024-05-23 08:09:54 -0400
  • 432b070bde Fix transcriptions endpoint Artificiangel 2024-05-23 08:07:51 -0400
  • ad54d524f7 Revert "Fix stopping strings for llama-3 and phi (#6043)" oobabooga 2024-05-22 17:18:08 -0700
  • 7c03e4af4d Improve DRY sampler performance 1 of 2 (simple data type changes). Belladore 2024-05-23 00:54:36 +0300
  • 5499bc9bc8
    Fix stopping strings for llama-3 and phi (#6043) oobabooga 2024-05-22 13:53:59 -0300
  • bfb9d7fc01 Simplify oobabooga 2024-05-22 09:44:14 -0700
  • ce89dbc5ff Account for llama.cpp loader oobabooga 2024-05-22 09:42:33 -0700
  • ee6728640b Minor fixes oobabooga 2024-05-22 08:24:16 -0700
  • 0730c191da Fix stopping strings for llama-3 and phi oobabooga 2024-05-22 08:18:02 -0700
  • a4df993d41
    Update README.md Shixian Sheng 2024-05-22 07:32:09 -0400
  • 52d74bb0d8
    Update README.md Shixian Sheng 2024-05-22 07:27:36 -0400
  • 68c46970d5
    Update README.md Shixian Sheng 2024-05-22 07:27:01 -0400
  • 39db799bf1 Added UnicodeDecodeError workaround for modules/llamacpp_model.py nclok1405 2024-05-22 09:54:35 +0900
  • 8aaa0a6f4e
    Fixed minor typo in docs - Training Tab.md (#6038) rohitanshu 2024-05-21 23:22:22 +0530
  • 9e189947d1 Minor fix after bd7cc4234d (thanks @belladoreai) oobabooga 2024-05-21 10:37:30 -0700
  • ae86292159 Fix getting Phi-3-small-128k-instruct logits oobabooga 2024-05-21 10:35:00 -0700
  • 21a4ac0845
    Fixed minor typo in docs - Training Tab.md rohitanshu 2024-05-21 22:59:47 +0530
  • bd7cc4234d
    Backend cleanup (#6025) oobabooga 2024-05-21 13:32:02 -0300
  • 5dba2743ff Merge branch 'dev' into backend-cleanup oobabooga 2024-05-21 09:29:14 -0700
  • e2fe90df04 Add chat-templates to .gitignore Touch-Night 2024-05-22 00:06:35 +0800
  • 690ebc533b Save chat template Touch-Night 2024-05-21 23:57:08 +0800
  • de33a036c9 clarify workaround comment marcel 2024-05-21 17:12:26 +0200
  • 9d85533240
    --- updated-dependencies: - dependency-name: optimum dependency-type: direct:production ... dependabot[bot] 2024-05-20 20:25:01 +0000
  • ab1c451efb
    --- updated-dependencies: - dependency-name: gradio dependency-type: direct:production ... dependabot[bot] 2024-05-20 20:24:56 +0000
  • 87125d5833
    --- updated-dependencies: - dependency-name: transformers dependency-type: direct:production ... dependabot[bot] 2024-05-20 20:24:46 +0000
  • b30e830230
    --- updated-dependencies: - dependency-name: hqq dependency-type: direct:production update-type: version-update:semver-patch ... dependabot[bot] 2024-05-20 20:24:41 +0000
  • ae7a644c8f
    --- updated-dependencies: - dependency-name: peft dependency-type: direct:production ... dependabot[bot] 2024-05-20 20:24:31 +0000
  • 6a1682aa95 README: update command-line flags with raw --help output oobabooga 2024-05-19 20:28:21 -0700
  • 852c943769
    DRY: A modern repetition penalty that reliably prevents looping (#5677) Philipp Emanuel Weidmann 2024-05-20 08:23:47 +0530
  • 6177676302 Merge branch 'dev' into p-e-w-dry oobabooga 2024-05-19 19:29:53 -0700
  • 9f77ed1b98
    --idle-timeout flag to unload the model if unused for N minutes (#6026) oobabooga 2024-05-19 23:29:39 -0300
  • 4bc923e29c Organize a bit (put close to no_repeat_ngram_size in the codebase) oobabooga 2024-05-19 19:26:15 -0700
  • 818b4e0354
    Let grammar escape backslashes (#5865) altoiddealer 2024-05-19 19:26:09 -0400
  • 907702c204
    Fix gguf multipart file loading (#5857) Tisjwlf 2024-05-20 01:22:09 +0200
  • d7bd3da35e
    Add Llama 3 instruction template (#5891) Guanghua Lu 2024-05-20 07:17:26 +0800
  • 255dfb2ead Prevent llama-3 derivatives from having wrong template oobabooga 2024-05-19 16:16:29 -0700
  • 5cb59707f3
    fix: grammar not support utf-8 (#5900) A0nameless0man 2024-05-20 07:10:39 +0800
  • 8456d13349
    [docs] small docker changes (#5917) Jari Van Melckebeke 2024-05-20 01:09:37 +0200
  • e10c779eb9 Revert README changes oobabooga 2024-05-19 16:08:25 -0700
  • b63dc4e325
    UI: Warn user if they are trying to load a model from no path (#6006) Samuel Wein 2024-05-20 01:05:17 +0200
  • 2de586f586
    Update accelerate requirement from ==0.27.* to ==0.30.* (#5989) dependabot[bot] 2024-05-19 20:03:18 -0300
  • 6b546a2c8b
    llama.cpp: increase the max threads from 32 to 256 (#5889) chr 2024-05-20 07:02:19 +0800
  • 8519e56aa4
    Bump sse-starlette from 1.6.5 to 2.1.0 dependabot[bot] 2024-05-19 22:02:05 +0000
  • abe5ddc883
    Merge pull request #6027 from oobabooga/dev oobabooga 2024-05-19 19:01:11 -0300
  • d35d92d8f6 Clean up oobabooga 2024-05-19 14:37:02 -0700
  • 2e43780189 --idle-timeout flag to unload the model if unused for N minutes oobabooga 2024-05-19 14:33:26 -0700
  • 13f37b79af
    Merge 49087d67a2 into a38a37b3b3 Nick Potafiy 2024-05-19 16:09:15 -0400
  • a38a37b3b3 llama.cpp: default n_gpu_layers to the maximum value for the model automatically oobabooga 2024-05-19 10:57:42 -0700
  • 0d7e5033f1 Remove obsolete code oobabooga 2024-05-19 09:05:57 -0700
  • 2d2b9308e1 Attempt to fix HQQ oobabooga 2024-05-19 09:00:07 -0700
  • dbb1094b0d Remove GPTQ-for-LLaMa oobabooga 2024-05-19 08:51:32 -0700
  • 89fe454878 Remove inject_fused_attention option from AutoGPTQ to make it functional oobabooga 2024-05-19 07:42:57 -0700
  • a3f8adc9b8 Bump transformers to 4.41 oobabooga 2024-05-19 07:23:28 -0700
  • 569f6b691b Bump AQLM to 1.1.5 oobabooga 2024-05-19 07:22:58 -0700
  • fbc63b39c3 Bump HQQ to 0.1.7.post2 oobabooga 2024-05-19 07:21:42 -0700
  • 0b21656256 Bump AutoGPTQ to 0.7.1, remove ROCm support, remove CUDA 11.8 support oobabooga 2024-05-19 07:20:05 -0700
  • 23928783fc Bump AutoAWQ to 0.2.5 oobabooga 2024-05-19 07:10:30 -0700
  • 6a3f97db61 Remove QuIP# oobabooga 2024-05-19 07:07:28 -0700
  • a4611232b7 Make --verbose output less spammy oobabooga 2024-05-18 09:57:00 -0700
  • 0d90b3a25c Bump llama-cpp-python to 0.2.75 oobabooga 2024-05-18 05:26:26 -0700
  • 7348482ed8
    Create wrangler.toml N3XT-G3N-DEV 2024-05-17 07:45:49 +0200
  • fee7122a58 increase the max threads from 32 to 256 hongruichen 2024-02-20 10:36:02 +0800
  • f9b2ff1616
    Merge branch 'oobabooga:dev' into dev Artificiangel 2024-05-15 06:27:21 -0400
  • 2585d07e2a
    Document DRY parameters in UI Philipp Emanuel Weidmann 2024-05-14 18:19:11 +0530
  • 80cbb8f230
    Support both JSON array notation and comma-separated strings for sequence breakers Philipp Emanuel Weidmann 2024-05-14 17:52:43 +0530
  • fc58c3b0ba
    Bring parameter order in constructor in line with rest of code Philipp Emanuel Weidmann 2024-05-14 17:20:08 +0530
  • 61f0c65eb3
    Update gradio requirement from ==4.26.* to ==4.31.* dependabot[bot] 2024-05-13 20:12:54 +0000
  • 8422027848
    Update api.py Raunak-Kumar7 2024-05-13 12:57:41 +0530
  • d1833271db
    Warn user if they are trying to load a model from no path Samuel Wein 2024-05-12 20:03:30 +0200
  • e225b0b995 downloader: fix downloading 01-ai/Yi-1.5-34B-Chat oobabooga 2024-05-12 10:43:50 -0700
  • 9557f49f2f Bump llama-cpp-python to 0.2.73 oobabooga 2024-05-11 10:53:19 -0700
  • a6420c444c Chat: fix HTML changes observing in main.js Corneliu Zuzu 2024-05-09 16:26:18 +0300
  • 4eb64a3716 Remove redundant check from js_chat_html_update Corneliu Zuzu 2024-05-09 14:37:40 +0300
  • 49087d67a2
    Merge branch 'oobabooga:main' into main Nick Potafiy 2024-05-09 00:00:31 -0400
  • fd8038dd8c Add Llama-v3 template to config.yaml Touch-Night 2024-05-09 10:35:51 +0800
  • 396753350a
    Merge branch 'oobabooga:main' into main Guanghua Lu 2024-05-09 10:21:38 +0800
  • 8ad2b6585a add additional record button next to "Generate" button marcel 2024-05-09 02:36:43 +0200
  • 9ac528715c
    Merge pull request #5996 from oobabooga/dev oobabooga 2024-05-08 16:37:26 -0300
  • 0a5259b926 Chat-mode only - add output_modifier_stream extensions handler! Corneliu Zuzu 2024-05-08 11:03:03 +0300
  • 75335c0987 Optimize chat HTML updates: handle structurally with JS Corneliu Zuzu 2024-05-08 10:58:16 +0300
  • ec716a308e Misc: add llama3 instruction template Corneliu Zuzu 2024-05-08 10:56:07 +0300
  • 2c43e02d8e Misc: shorthand for enabling remote debugging Corneliu Zuzu 2024-05-08 10:55:13 +0300
  • 7a728a38eb Update README oobabooga 2024-05-07 02:59:36 -0700
  • d5bde7babc UI: improve the performance of code syntax highlighting oobabooga 2024-05-06 16:51:05 -0700
  • 093b31ed8b
    Update accelerate requirement from ==0.27.* to ==0.30.* dependabot[bot] 2024-05-06 20:56:42 +0000
  • a01dc2fafd
    Update gradio requirement from ==4.26.* to ==4.29.* dependabot[bot] 2024-05-06 20:55:56 +0000
  • 1237137028 Updated Llama-3 instruction template Nick Potafiy 2024-05-04 22:34:06 -0400
  • 0b193b8553 Downloader: handle one more retry case after 5770e06c48 oobabooga 2024-05-04 19:25:04 -0700
  • 4f33759c9c Updated Llama-3 instruction template Nick Potafiy 2024-05-04 16:52:57 -0400
  • 95f5263b70 Added Llama-3 instruction template Nick Potafiy 2024-05-04 16:17:00 -0400
  • b7a5bc6356 Zero memorey must be integer Nick Potafiy 2024-05-04 14:11:38 -0400
  • 7f4c486fc1 updating the correct map Nick Potafiy 2024-05-04 13:59:43 -0400
  • 0dfb7c34d1 Regex pattern more restrictive Nick Potafiy 2024-05-04 12:46:03 -0400
  • 88a1383cf9 Fixed issue with allocating 0 max_memory to GPU Nick Potafiy 2024-05-04 12:28:10 -0400
  • f0332c07be Checks & try block added, returns last tokens hidden state Nick Potafiy 2024-05-04 02:35:57 -0400
  • cd700ebade Returning dict instead of array Nick Potafiy 2024-05-04 02:17:31 -0400
  • df4cf32622 Reverted Nick Potafiy 2024-05-04 01:58:10 -0400
  • b10fc0d87a last_hidden_state api endpoint Nick Potafiy 2024-05-04 01:53:22 -0400