Clear the cache for a specific model or the entire cache directory.
Parameters:
Name |
Type |
Description |
Default |
model_id
|
Optional[str]
|
The model ID to clear cache for. If None, clears entire cache. Defaults to None.
|
None
|
delete_from_disk
|
bool
|
Whether to delete cached files from disk. Defaults to False.
|
True
|
Source code in inference/core/cache/model_artifacts.py
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238 | def clear_cache(model_id: Optional[str] = None, delete_from_disk: bool = True) -> None:
"""Clear the cache for a specific model or the entire cache directory.
Args:
model_id (Optional[str], optional): The model ID to clear cache for. If None, clears entire cache. Defaults to None.
delete_from_disk (bool, optional): Whether to delete cached files from disk. Defaults to False.
"""
if not delete_from_disk:
return
cache_dir = get_cache_dir(model_id=model_id)
if not os.path.exists(cache_dir):
return
lock_dir = MODEL_CACHE_DIR + "/_file_locks" # Dedicated lock directory
os.makedirs(lock_dir, exist_ok=True) # ensure lock directory exists.
# Use the last 2 levels of the cache directory path as the lock file name suffix
parts = os.path.normpath(cache_dir).split(os.sep)
suffix = (
os.path.join(*parts[-2:]) if len(parts) >= 2 else os.path.basename(cache_dir)
)
lock_file = os.path.join(lock_dir, f"{suffix}.lock")
try:
lock = FileLock(lock_file, timeout=10) # 10 second timeout
with lock:
if not os.path.exists(cache_dir): # Check again after acquiring lock
return # Already deleted by another process
max_retries = 3
retry_delay = 1 # Initial delay in seconds
for attempt in range(max_retries):
try:
shutil.rmtree(cache_dir, onerror=_rmtree_onerror)
return # Success
except FileNotFoundError:
return # Already deleted by another process
except Exception as e:
if attempt < max_retries - 1:
logger.warning(
f"Error deleting cache %s: %s, retrying in %s seconds...",
cache_dir,
e,
retry_delay,
)
time.sleep(retry_delay)
retry_delay *= 2 # Exponential backoff
else:
logger.warning(
f"Error deleting cache %s: %s, max retries exceeded.",
cache_dir,
e,
)
return
except Exception as e:
logger.warning(
f"Error acquiring lock for cache %s, skipping cache cleanup. %s",
cache_dir,
e,
)
finally:
try:
if os.path.exists(lock_file):
os.unlink(lock_file) # Clean up lock file
except OSError:
pass # Best effort cleanup
|