10
10
import ast
11
11
import asyncio
12
12
import hashlib
13
+ import inspect
13
14
import io
14
15
import json
15
16
import logging
@@ -245,7 +246,8 @@ def __init__(
245
246
self .redis_flush_size = redis_flush_size
246
247
self .redis_version = "Unknown"
247
248
try :
248
- self .redis_version = self .redis_client .info ()["redis_version" ]
249
+ if not inspect .iscoroutinefunction (self .redis_client ):
250
+ self .redis_version = self .redis_client .info ()["redis_version" ] # type: ignore
249
251
except Exception :
250
252
pass
251
253
@@ -266,7 +268,8 @@ def __init__(
266
268
267
269
### SYNC HEALTH PING ###
268
270
try :
269
- self .redis_client .ping ()
271
+ if hasattr (self .redis_client , "ping" ):
272
+ self .redis_client .ping () # type: ignore
270
273
except Exception as e :
271
274
verbose_logger .error (
272
275
"Error connecting to Sync Redis client" , extra = {"error" : str (e )}
@@ -308,7 +311,7 @@ def increment_cache(
308
311
_redis_client = self .redis_client
309
312
start_time = time .time ()
310
313
try :
311
- result = _redis_client .incr (name = key , amount = value )
314
+ result : int = _redis_client .incr (name = key , amount = value ) # type: ignore
312
315
313
316
if ttl is not None :
314
317
# check if key already has ttl, if not -> set ttl
@@ -561,7 +564,7 @@ async def async_set_cache_sadd(
561
564
f"Set ASYNC Redis Cache: key: { key } \n Value { value } \n ttl={ ttl } "
562
565
)
563
566
try :
564
- await redis_client .sadd (key , * value )
567
+ await redis_client .sadd (key , * value ) # type: ignore
565
568
if ttl is not None :
566
569
_td = timedelta (seconds = ttl )
567
570
await redis_client .expire (key , _td )
@@ -712,7 +715,7 @@ def batch_get_cache(self, key_list) -> dict:
712
715
for cache_key in key_list :
713
716
cache_key = self .check_and_fix_namespace (key = cache_key )
714
717
_keys .append (cache_key )
715
- results = self .redis_client .mget (keys = _keys )
718
+ results : List = self .redis_client .mget (keys = _keys ) # type: ignore
716
719
717
720
# Associate the results back with their keys.
718
721
# 'results' is a list of values corresponding to the order of keys in 'key_list'.
@@ -842,7 +845,7 @@ def sync_ping(self) -> bool:
842
845
print_verbose ("Pinging Sync Redis Cache" )
843
846
start_time = time .time ()
844
847
try :
845
- response = self .redis_client .ping ()
848
+ response : bool = self .redis_client .ping () # type: ignore
846
849
print_verbose (f"Redis Cache PING: { response } " )
847
850
## LOGGING ##
848
851
end_time = time .time ()
@@ -911,8 +914,8 @@ async def delete_cache_keys(self, keys):
911
914
async with _redis_client as redis_client :
912
915
await redis_client .delete (* keys )
913
916
914
- def client_list (self ):
915
- client_list = self .redis_client .client_list ()
917
+ def client_list (self ) -> List :
918
+ client_list : List = self .redis_client .client_list () # type: ignore
916
919
return client_list
917
920
918
921
def info (self ):
0 commit comments