Skip to content

Commit ae01d33

Browse files
committed
Updated Version
1 parent b1c7d10 commit ae01d33

File tree

3 files changed

+35
-2
lines changed

3 files changed

+35
-2
lines changed

CHANGELOG.md

Lines changed: 33 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -2,6 +2,39 @@
22

33
All notable changes to LocalLab will be documented in this file.
44

5+
## [0.7.0] - 2025-05-16
6+
7+
### Improved
8+
9+
- Significantly enhanced stream generation quality with comprehensive improvements:
10+
11+
- Improved token generation parameters for higher quality responses
12+
- Enhanced stop sequence detection with better conversation markers handling
13+
- Implemented more intelligent repetition detection to prevent loops
14+
- Optimized token buffering and yielding logic for smoother streaming
15+
- Added better error handling and recovery in streaming responses
16+
17+
- Improved non-streaming generation quality across all endpoints:
18+
19+
- Enhanced generate, chat, and batch generation methods with optimized parameters
20+
- Implemented repetition detection to prevent the model from getting stuck
21+
- Added comprehensive special token handling and cleanup
22+
- Improved conversation marker detection for better response termination
23+
- Balanced parameters between quality and speed for optimal performance
24+
25+
- Optimized memory management:
26+
- Reduced frequency of memory checks to avoid interrupting generation
27+
- Implemented smarter memory threshold for cache clearing
28+
- Added better error recovery for out-of-memory situations
29+
30+
### Changed
31+
32+
- Increased default max_length from 2048 to 4096 for non-streaming generation
33+
- Increased token generation batch size from 4 to 8 for better efficiency
34+
- Adjusted top_k (80), top_p (0.92), and repetition_penalty (1.15) for better quality
35+
- Increased max_time parameter to 180 seconds for more complete responses
36+
- Enhanced all generation endpoints with consistent high-quality parameters
37+
538
## [0.6.6] - 2025-05-16
639

740
### Fixed

locallab/__init__.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@
66
# This ensures Hugging Face's progress bars are displayed correctly
77
from .utils.early_config import configure_hf_logging
88

9-
__version__ = "0.6.6" # Fixed Hugging Face progress bars display with version-agnostic approach
9+
__version__ = "0.7.0" # Improved stream generation and non-streaming generation quality
1010

1111
# Only import what's necessary initially, lazy-load the rest
1212
from .logger import get_logger

setup.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -47,7 +47,7 @@
4747

4848
setup(
4949
name="locallab",
50-
version="0.6.6",
50+
version="0.7.0",
5151
packages=find_packages(include=["locallab", "locallab.*"]),
5252
install_requires=install_requires,
5353
extras_require={

0 commit comments

Comments
 (0)