Releases: containers/ramalama
Releases · containers/ramalama
v0.5.2
What's Changed
- This is all dead code which isn't called by @ericcurtin in #574
- On ARM by default turn on GPU acceleration by @ericcurtin in #573
- Capitalize constants in python files (CONSTANT_CASE) by @swarajpande5 in #579
- Add flake by @jim3692 in #581
- Update llama.cpp to include minor llama-run by @ericcurtin in #580
- Simplify this comparison by @ericcurtin in #576
- Fix ramalama run on docker to work correctly by @rhatdan in #583
- granite-code models in Ollama are malformed by @ericcurtin in #584
- Bump to v0.5.2 by @rhatdan in #585
New Contributors
Full Changelog: v0.5.1...v0.5.2
v0.5.1
What's Changed
- When it debug mode show the exception for subcommand missing by @rhatdan in #568
- Revert "If podman machine is running use containers" by @ericcurtin in #570
- Check all files are adding to packaging by @ericcurtin in #569
- Add podman info or docker info to ramalama info output by @rhatdan in #567
- Bump to v0.5.1 by @rhatdan in #571
Full Changelog: v0.5.0...v0.5.1
v0.5.0
What's Changed
- Remove some more gfx9 stuff by @ericcurtin in #554
- Only use the X/Y parts of the version when pulling images by @rhatdan in #555
- If podman machine is running use containers by @ericcurtin in #557
- Update llama.cpp to contain llama-run fix by @ericcurtin in #556
- We are passing the same value in twice, causes exception by @rhatdan in #558
- Remove firefox from builds by @ericcurtin in #561
- Remove cruft added to redirect llama-cli 2>/dev/null by @rhatdan in #559
- run_cmd to exec_cmd by @ericcurtin in #564
- smollm:135m for testing purposes by @ericcurtin in #563
- Fix podman stop --all by @rhatdan in #562
- Bump to v0.5.0 by @rhatdan in #565
Full Changelog: v0.4.0...v0.5.0
v0.4.0
What's Changed
- ramalama rm should require at least one argument by @rhatdan in #516
- Fix handling of missing args for login/logout by @rhatdan in #517
- Add --version and fix sort order of commands by @rhatdan in #518
- Update shortnames.conf to alias new granite models by @ericcurtin in #519
- Prettify and improve download bars during model pulls by @swarajpande5 in #521
- Remove
tqdm
dependency by @swarajpande5 in #523 - Relocate
File
andHttpClient
to their respective modules by @swarajpande5 in #522 - Add missing files to install.sh script by @ericcurtin in #524
- Change pip to pipx in installation instructions by @ericcurtin in #520
- Pass ASAHI, CUDA, HIP, HSA prefixed env vars to container by @abn in #526
- Resolves - Debug option shows traceback when no subcommand is passed by @pghole in #527
- Skip codespell of
venv
by adding--skip
in the command by @swarajpande5 in #528 - Improved command handling and refactoring by @ericcurtin in #531
- Update README and model.py for Nvidia GPU support by @ericcurtin in #530
- Add UI support documentation by @ericcurtin in #529
- Add llama-run by @ericcurtin in #452
- Update container images to use latest by @ericcurtin in #534
- Update llama.cpp to get context size fix by @ericcurtin in #538
- If runtime is vllm, use image from modh quay.io by @ericcurtin in #533
- Remove huggingface warning by @ericcurtin in #540
- When using podman run add --pull=newer to update container image by @rhatdan in #541
- Build to build_rm by @ericcurtin in #543
- whisper.cpp now installs properly by @rhatdan in #544
- Re-add mkdir -p by @ericcurtin in #545
- This might temporarily fix an issue we are having by @ericcurtin in #547
- Attempt to use versioned container by @ericcurtin in #537
- Fix llama-run to be installed with correct library paths. by @rhatdan in #549
- Add CPU inferencing container build steps to CI by @ericcurtin in #551
- This should save some space by @ericcurtin in #552
- Bump to v0.4.0 by @rhatdan in #550
New Contributors
Full Changelog: v0.3.0...v0.4.0
v0.3.0
What's Changed
- Move man page README.md to full links by @rhatdan in #483
- Allow users to set ctx-size via command line by @rhatdan in #484
- Add --seed option by @rhatdan in #485
- Update install.sh by @jtligon in #493
- Take in fixes from @Churchyard to modernize spec file by @smooge in #494
- Fix up building and pushing OCI Images by @rhatdan in #492
- Fix handling of file_not_found errors by @rhatdan in #499
- Updated nv docs to align with latest WSL2 cuda setup by @bmahabirbu in #503
- Add ramalama convert command by @rhatdan in #500
- Stop checking if command is running in container by @rhatdan in #505
- Add initial CONTRIBUTING.md file by @rhatdan in #507
- Place image name just before command by @ericcurtin in #511
- Simplify install by @ericcurtin in #510
- Fix handling of README.md in docs directory by @rhatdan in #512
- Add installation steps for Podman 5 in CI workflows by @ericcurtin in #508
- Bump to v0.3.0 by @rhatdan in #513
New Contributors
Full Changelog: v0.2.0...v0.3.0
v0.2.0
v0.1.3
What's Changed
- Enable GCC Toolet 12 to support AVX VNNI by @nzwulfin in #473
- Failover to OCI when push fails with default push mechanism by @rhatdan in #476
- Fall back to huggingface-cli when pulling via URL fails by @rhatdan in #475
- Revert "Switch to llama-simple-chat" by @rhatdan in #477
- Add support for http, https and file pulls by @rhatdan in #463
- Bump to v0.1.3 by @rhatdan in #479
New Contributors
Full Changelog: v0.1.2...v0.1.3
v0.1.2
What's Changed
- Bump to v0.1.1 by @rhatdan in #450
- Update ggerganov/whisper.cpp digest to f19463e by @renovate in #453
- Switch to llama-simple-chat by @ericcurtin in #454
- Simplify container image build by @ericcurtin in #451
- Update ggerganov/whisper.cpp digest to 83ac284 by @renovate in #455
- cli.py: remove errant slash preventing the loading of user conf file(s) by @FNGarvin in #457
- Update ggerganov/whisper.cpp digest to f02b40b by @renovate in #456
- Switched DGGML_CUDA to ON in cuda containerfile by @bmahabirbu in #459
- Update ggerganov/whisper.cpp digest to bb12cd9 by @renovate in #460
- Update ggerganov/whisper.cpp digest to 01d3bd7 by @renovate in #461
- Update ggerganov/whisper.cpp digest to d24f981 by @renovate in #462
- Docu by @atarlov in #464
- Update ggerganov/whisper.cpp digest to 6266a9f by @renovate in #466
- Fix handling of ramalama login huggingface by @rhatdan in #467
- Support huggingface-cli older than 0.25.0, like on Fedora 40 and 41 by @debarshiray in #468
- Bump to v0.1.2 by @rhatdan in #470
New Contributors
- @FNGarvin made their first contribution in #457
- @atarlov made their first contribution in #464
- @debarshiray made their first contribution in #468
Full Changelog: v0.1.1...v0.1.2
v0.1.1
Full Changelog: v0.1.0...v0.1.1
Mainly to fix issue in PyPi
v0.1.0
What's Changed
- We can now run models via Kompute in podman-machine by @ericcurtin in #440
- Only do dnf install for cuda images by @ericcurtin in #441
- Add --host=0.0.0.0 if running llama.cpp serve within a container by @rhatdan in #444
- Document the host flag in ramalama.conf file by @rhatdan in #447
- Add granite-8b to shortnames.conf by @rhatdan in #448
- Fix RamaLama container image build by @ericcurtin in #446
- Bump to v0.1.0 by @rhatdan in #449
Full Changelog: v0.0.23...v0.1.0