@@ -42,21 +42,21 @@ def __init__(self, conf, greedy=True):
42
42
s3 .get_connection (conf )
43
43
)
44
44
45
- self ._bucket_name_measures = (
45
+ self ._bucket_name = (
46
46
self ._bucket_prefix + "-" + self .MEASURE_PREFIX
47
47
)
48
48
49
49
def __str__ (self ):
50
- return "%s: %s" % (self .__class__ .__name__ , self ._bucket_name_measures )
50
+ return "%s: %s" % (self .__class__ .__name__ , self ._bucket_name )
51
51
52
52
def _get_storage_sacks (self ):
53
- response = self .s3 .get_object (Bucket = self ._bucket_name_measures ,
53
+ response = self .s3 .get_object (Bucket = self ._bucket_name ,
54
54
Key = self .CFG_PREFIX )
55
55
return json .loads (response ['Body' ].read ().decode ())[self .CFG_SACKS ]
56
56
57
57
def set_storage_settings (self , num_sacks ):
58
58
data = {self .CFG_SACKS : num_sacks }
59
- self .s3 .put_object (Bucket = self ._bucket_name_measures ,
59
+ self .s3 .put_object (Bucket = self ._bucket_name ,
60
60
Key = self .CFG_PREFIX ,
61
61
Body = json .dumps (data ).encode ())
62
62
@@ -67,7 +67,7 @@ def remove_sacks(num_sacks):
67
67
68
68
def upgrade (self , num_sacks ):
69
69
try :
70
- s3 .create_bucket (self .s3 , self ._bucket_name_measures ,
70
+ s3 .create_bucket (self .s3 , self ._bucket_name ,
71
71
self ._region_name )
72
72
except botocore .exceptions .ClientError as e :
73
73
if e .response ['Error' ].get ('Code' ) not in (
@@ -80,7 +80,7 @@ def upgrade(self, num_sacks):
80
80
def _store_new_measures (self , metric_id , data ):
81
81
now = datetime .datetime .utcnow ().strftime ("_%Y%m%d_%H:%M:%S" )
82
82
self .s3 .put_object (
83
- Bucket = self ._bucket_name_measures ,
83
+ Bucket = self ._bucket_name ,
84
84
Key = "/" .join ((str (self .sack_for_metric (metric_id )),
85
85
str (metric_id ),
86
86
str (uuid .uuid4 ()) + now )),
@@ -97,7 +97,7 @@ def _build_report(self, details):
97
97
else :
98
98
kwargs = {}
99
99
response = self .s3 .list_objects_v2 (
100
- Bucket = self ._bucket_name_measures ,
100
+ Bucket = self ._bucket_name ,
101
101
** kwargs )
102
102
for c in response .get ('Contents' , ()):
103
103
if c ['Key' ] != self .CFG_PREFIX :
@@ -118,7 +118,7 @@ def _list_files(self, path_items, **kwargs):
118
118
except KeyError :
119
119
pass
120
120
response = self .s3 .list_objects_v2 (
121
- Bucket = self ._bucket_name_measures ,
121
+ Bucket = self ._bucket_name ,
122
122
Prefix = "/" .join (path_items ) + "/" ,
123
123
** kwargs )
124
124
yield response
@@ -136,7 +136,7 @@ def _list_measure_files_for_metric(self, sack, metric_id):
136
136
def delete_unprocessed_measures_for_metric (self , metric_id ):
137
137
sack = self .sack_for_metric (metric_id )
138
138
files = self ._list_measure_files_for_metric (sack , metric_id )
139
- s3 .bulk_delete (self .s3 , self ._bucket_name_measures , files )
139
+ s3 .bulk_delete (self .s3 , self ._bucket_name , files )
140
140
141
141
def has_unprocessed (self , metric_id ):
142
142
sack = self .sack_for_metric (metric_id )
@@ -152,7 +152,7 @@ def process_measure_for_metrics(self, metric_ids):
152
152
all_files .extend (files )
153
153
for f in files :
154
154
response = self .s3 .get_object (
155
- Bucket = self ._bucket_name_measures ,
155
+ Bucket = self ._bucket_name ,
156
156
Key = f )
157
157
measures [metric_id ] = numpy .concatenate ((
158
158
measures [metric_id ],
@@ -162,7 +162,7 @@ def process_measure_for_metrics(self, metric_ids):
162
162
yield measures
163
163
164
164
# Now clean objects
165
- s3 .bulk_delete (self .s3 , self ._bucket_name_measures , all_files )
165
+ s3 .bulk_delete (self .s3 , self ._bucket_name , all_files )
166
166
167
167
@contextlib .contextmanager
168
168
def process_measures_for_sack (self , sack ):
@@ -177,7 +177,7 @@ def process_measures_for_sack(self, sack):
177
177
continue
178
178
179
179
response = self .s3 .get_object (
180
- Bucket = self ._bucket_name_measures ,
180
+ Bucket = self ._bucket_name ,
181
181
Key = f )
182
182
measures [metric_id ] = numpy .concatenate ((
183
183
measures [metric_id ],
@@ -187,4 +187,6 @@ def process_measures_for_sack(self, sack):
187
187
yield measures
188
188
189
189
# Now clean objects
190
- s3 .bulk_delete (self .s3 , self ._bucket_name_measures , files )
190
+ s3 .bulk_delete (self .s3 , self ._bucket_name , files )
191
+
192
+ get_health_status = s3 .get_s3_health_status
0 commit comments