xref: /linux/fs/ceph/metric.c (revision e467705a9fb37f51595aa6deaca085ccb4005454)
1 /* SPDX-License-Identifier: GPL-2.0 */
2 #include <linux/ceph/ceph_debug.h>
3 
4 #include <linux/types.h>
5 #include <linux/percpu_counter.h>
6 #include <linux/math64.h>
7 
8 #include "metric.h"
9 #include "mds_client.h"
10 
11 static void ktime_to_ceph_timespec(struct ceph_timespec *ts, ktime_t val)
12 {
13 	struct timespec64 t = ktime_to_timespec64(val);
14 	ceph_encode_timespec64(ts, &t);
15 }
16 
17 static bool ceph_mdsc_send_metrics(struct ceph_mds_client *mdsc,
18 				   struct ceph_mds_session *s)
19 {
20 	struct ceph_metric_head *head;
21 	struct ceph_metric_cap *cap;
22 	struct ceph_metric_read_latency *read;
23 	struct ceph_metric_write_latency *write;
24 	struct ceph_metric_metadata_latency *meta;
25 	struct ceph_metric_dlease *dlease;
26 	struct ceph_opened_files *files;
27 	struct ceph_pinned_icaps *icaps;
28 	struct ceph_opened_inodes *inodes;
29 	struct ceph_read_io_size *rsize;
30 	struct ceph_write_io_size *wsize;
31 	struct ceph_client_metric *m = &mdsc->metric;
32 	u64 nr_caps = atomic64_read(&m->total_caps);
33 	u32 header_len = sizeof(struct ceph_metric_header);
34 	struct ceph_client *cl = mdsc->fsc->client;
35 	struct ceph_msg *msg;
36 	s64 sum;
37 	s32 items = 0;
38 	s32 len;
39 
40 	/* Do not send the metrics until the MDS rank is ready */
41 	mutex_lock(&mdsc->mutex);
42 	if (ceph_mdsmap_get_state(mdsc->mdsmap, s->s_mds) != CEPH_MDS_STATE_ACTIVE) {
43 		mutex_unlock(&mdsc->mutex);
44 		return false;
45 	}
46 	mutex_unlock(&mdsc->mutex);
47 
48 	len = sizeof(*head) + sizeof(*cap) + sizeof(*read) + sizeof(*write)
49 	      + sizeof(*meta) + sizeof(*dlease) + sizeof(*files)
50 	      + sizeof(*icaps) + sizeof(*inodes) + sizeof(*rsize)
51 	      + sizeof(*wsize);
52 
53 	msg = ceph_msg_new(CEPH_MSG_CLIENT_METRICS, len, GFP_NOFS, true);
54 	if (!msg) {
55 		pr_err_client(cl, "to mds%d, failed to allocate message\n",
56 			      s->s_mds);
57 		return false;
58 	}
59 
60 	head = msg->front.iov_base;
61 
62 	/* encode the cap metric */
63 	cap = (struct ceph_metric_cap *)(head + 1);
64 	cap->header.type = cpu_to_le32(CLIENT_METRIC_TYPE_CAP_INFO);
65 	cap->header.ver = 1;
66 	cap->header.compat = 1;
67 	cap->header.data_len = cpu_to_le32(sizeof(*cap) - header_len);
68 	cap->hit = cpu_to_le64(percpu_counter_sum(&m->i_caps_hit));
69 	cap->mis = cpu_to_le64(percpu_counter_sum(&m->i_caps_mis));
70 	cap->total = cpu_to_le64(nr_caps);
71 	items++;
72 
73 	/* encode the read latency metric */
74 	read = (struct ceph_metric_read_latency *)(cap + 1);
75 	read->header.type = cpu_to_le32(CLIENT_METRIC_TYPE_READ_LATENCY);
76 	read->header.ver = 2;
77 	read->header.compat = 1;
78 	read->header.data_len = cpu_to_le32(sizeof(*read) - header_len);
79 	sum = m->metric[METRIC_READ].latency_sum;
80 	ktime_to_ceph_timespec(&read->lat, sum);
81 	ktime_to_ceph_timespec(&read->avg, m->metric[METRIC_READ].latency_avg);
82 	read->sq_sum = cpu_to_le64(m->metric[METRIC_READ].latency_sq_sum);
83 	read->count = cpu_to_le64(m->metric[METRIC_READ].total);
84 	items++;
85 
86 	/* encode the write latency metric */
87 	write = (struct ceph_metric_write_latency *)(read + 1);
88 	write->header.type = cpu_to_le32(CLIENT_METRIC_TYPE_WRITE_LATENCY);
89 	write->header.ver = 2;
90 	write->header.compat = 1;
91 	write->header.data_len = cpu_to_le32(sizeof(*write) - header_len);
92 	sum = m->metric[METRIC_WRITE].latency_sum;
93 	ktime_to_ceph_timespec(&write->lat, sum);
94 	ktime_to_ceph_timespec(&write->avg, m->metric[METRIC_WRITE].latency_avg);
95 	write->sq_sum = cpu_to_le64(m->metric[METRIC_WRITE].latency_sq_sum);
96 	write->count = cpu_to_le64(m->metric[METRIC_WRITE].total);
97 	items++;
98 
99 	/* encode the metadata latency metric */
100 	meta = (struct ceph_metric_metadata_latency *)(write + 1);
101 	meta->header.type = cpu_to_le32(CLIENT_METRIC_TYPE_METADATA_LATENCY);
102 	meta->header.ver = 2;
103 	meta->header.compat = 1;
104 	meta->header.data_len = cpu_to_le32(sizeof(*meta) - header_len);
105 	sum = m->metric[METRIC_METADATA].latency_sum;
106 	ktime_to_ceph_timespec(&meta->lat, sum);
107 	ktime_to_ceph_timespec(&meta->avg, m->metric[METRIC_METADATA].latency_avg);
108 	meta->sq_sum = cpu_to_le64(m->metric[METRIC_METADATA].latency_sq_sum);
109 	meta->count = cpu_to_le64(m->metric[METRIC_METADATA].total);
110 	items++;
111 
112 	/* encode the dentry lease metric */
113 	dlease = (struct ceph_metric_dlease *)(meta + 1);
114 	dlease->header.type = cpu_to_le32(CLIENT_METRIC_TYPE_DENTRY_LEASE);
115 	dlease->header.ver = 1;
116 	dlease->header.compat = 1;
117 	dlease->header.data_len = cpu_to_le32(sizeof(*dlease) - header_len);
118 	dlease->hit = cpu_to_le64(percpu_counter_sum(&m->d_lease_hit));
119 	dlease->mis = cpu_to_le64(percpu_counter_sum(&m->d_lease_mis));
120 	dlease->total = cpu_to_le64(atomic64_read(&m->total_dentries));
121 	items++;
122 
123 	sum = percpu_counter_sum(&m->total_inodes);
124 
125 	/* encode the opened files metric */
126 	files = (struct ceph_opened_files *)(dlease + 1);
127 	files->header.type = cpu_to_le32(CLIENT_METRIC_TYPE_OPENED_FILES);
128 	files->header.ver = 1;
129 	files->header.compat = 1;
130 	files->header.data_len = cpu_to_le32(sizeof(*files) - header_len);
131 	files->opened_files = cpu_to_le64(atomic64_read(&m->opened_files));
132 	files->total = cpu_to_le64(sum);
133 	items++;
134 
135 	/* encode the pinned icaps metric */
136 	icaps = (struct ceph_pinned_icaps *)(files + 1);
137 	icaps->header.type = cpu_to_le32(CLIENT_METRIC_TYPE_PINNED_ICAPS);
138 	icaps->header.ver = 1;
139 	icaps->header.compat = 1;
140 	icaps->header.data_len = cpu_to_le32(sizeof(*icaps) - header_len);
141 	icaps->pinned_icaps = cpu_to_le64(nr_caps);
142 	icaps->total = cpu_to_le64(sum);
143 	items++;
144 
145 	/* encode the opened inodes metric */
146 	inodes = (struct ceph_opened_inodes *)(icaps + 1);
147 	inodes->header.type = cpu_to_le32(CLIENT_METRIC_TYPE_OPENED_INODES);
148 	inodes->header.ver = 1;
149 	inodes->header.compat = 1;
150 	inodes->header.data_len = cpu_to_le32(sizeof(*inodes) - header_len);
151 	inodes->opened_inodes = cpu_to_le64(percpu_counter_sum(&m->opened_inodes));
152 	inodes->total = cpu_to_le64(sum);
153 	items++;
154 
155 	/* encode the read io size metric */
156 	rsize = (struct ceph_read_io_size *)(inodes + 1);
157 	rsize->header.type = cpu_to_le32(CLIENT_METRIC_TYPE_READ_IO_SIZES);
158 	rsize->header.ver = 1;
159 	rsize->header.compat = 1;
160 	rsize->header.data_len = cpu_to_le32(sizeof(*rsize) - header_len);
161 	rsize->total_ops = cpu_to_le64(m->metric[METRIC_READ].total);
162 	rsize->total_size = cpu_to_le64(m->metric[METRIC_READ].size_sum);
163 	items++;
164 
165 	/* encode the write io size metric */
166 	wsize = (struct ceph_write_io_size *)(rsize + 1);
167 	wsize->header.type = cpu_to_le32(CLIENT_METRIC_TYPE_WRITE_IO_SIZES);
168 	wsize->header.ver = 1;
169 	wsize->header.compat = 1;
170 	wsize->header.data_len = cpu_to_le32(sizeof(*wsize) - header_len);
171 	wsize->total_ops = cpu_to_le64(m->metric[METRIC_WRITE].total);
172 	wsize->total_size = cpu_to_le64(m->metric[METRIC_WRITE].size_sum);
173 	items++;
174 
175 	put_unaligned_le32(items, &head->num);
176 	msg->front.iov_len = len;
177 	msg->hdr.version = cpu_to_le16(1);
178 	msg->hdr.compat_version = cpu_to_le16(1);
179 	msg->hdr.front_len = cpu_to_le32(msg->front.iov_len);
180 	ceph_con_send(&s->s_con, msg);
181 
182 	return true;
183 }
184 
185 
186 static void metric_get_session(struct ceph_mds_client *mdsc)
187 {
188 	struct ceph_mds_session *s;
189 	int i;
190 
191 	mutex_lock(&mdsc->mutex);
192 	for (i = 0; i < mdsc->max_sessions; i++) {
193 		s = __ceph_lookup_mds_session(mdsc, i);
194 		if (!s)
195 			continue;
196 
197 		/*
198 		 * Skip it if MDS doesn't support the metric collection,
199 		 * or the MDS will close the session's socket connection
200 		 * directly when it get this message.
201 		 */
202 		if (check_session_state(s) &&
203 		    test_bit(CEPHFS_FEATURE_METRIC_COLLECT, &s->s_features)) {
204 			mdsc->metric.session = s;
205 			break;
206 		}
207 
208 		ceph_put_mds_session(s);
209 	}
210 	mutex_unlock(&mdsc->mutex);
211 }
212 
213 static void metric_delayed_work(struct work_struct *work)
214 {
215 	struct ceph_client_metric *m =
216 		container_of(work, struct ceph_client_metric, delayed_work.work);
217 	struct ceph_mds_client *mdsc =
218 		container_of(m, struct ceph_mds_client, metric);
219 
220 	if (mdsc->stopping || disable_send_metrics)
221 		return;
222 
223 	if (!m->session || !check_session_state(m->session)) {
224 		if (m->session) {
225 			ceph_put_mds_session(m->session);
226 			m->session = NULL;
227 		}
228 		metric_get_session(mdsc);
229 	}
230 	if (m->session) {
231 		ceph_mdsc_send_metrics(mdsc, m->session);
232 		metric_schedule_delayed(m);
233 	}
234 }
235 
236 int ceph_metric_init(struct ceph_client_metric *m)
237 {
238 	struct ceph_metric *metric;
239 	int ret, i;
240 
241 	if (!m)
242 		return -EINVAL;
243 
244 	atomic64_set(&m->total_dentries, 0);
245 	ret = percpu_counter_init(&m->d_lease_hit, 0, GFP_KERNEL);
246 	if (ret)
247 		return ret;
248 
249 	ret = percpu_counter_init(&m->d_lease_mis, 0, GFP_KERNEL);
250 	if (ret)
251 		goto err_d_lease_mis;
252 
253 	atomic64_set(&m->total_caps, 0);
254 	ret = percpu_counter_init(&m->i_caps_hit, 0, GFP_KERNEL);
255 	if (ret)
256 		goto err_i_caps_hit;
257 
258 	ret = percpu_counter_init(&m->i_caps_mis, 0, GFP_KERNEL);
259 	if (ret)
260 		goto err_i_caps_mis;
261 
262 	for (i = 0; i < METRIC_MAX; i++) {
263 		metric = &m->metric[i];
264 		spin_lock_init(&metric->lock);
265 		metric->size_sum = 0;
266 		metric->size_min = U64_MAX;
267 		metric->size_max = 0;
268 		metric->total = 0;
269 		metric->latency_sum = 0;
270 		metric->latency_avg = 0;
271 		metric->latency_sq_sum = 0;
272 		metric->latency_min = KTIME_MAX;
273 		metric->latency_max = 0;
274 	}
275 
276 	atomic64_set(&m->opened_files, 0);
277 	ret = percpu_counter_init(&m->opened_inodes, 0, GFP_KERNEL);
278 	if (ret)
279 		goto err_opened_inodes;
280 	ret = percpu_counter_init(&m->total_inodes, 0, GFP_KERNEL);
281 	if (ret)
282 		goto err_total_inodes;
283 
284 	m->session = NULL;
285 	INIT_DELAYED_WORK(&m->delayed_work, metric_delayed_work);
286 
287 	return 0;
288 
289 err_total_inodes:
290 	percpu_counter_destroy(&m->opened_inodes);
291 err_opened_inodes:
292 	percpu_counter_destroy(&m->i_caps_mis);
293 err_i_caps_mis:
294 	percpu_counter_destroy(&m->i_caps_hit);
295 err_i_caps_hit:
296 	percpu_counter_destroy(&m->d_lease_mis);
297 err_d_lease_mis:
298 	percpu_counter_destroy(&m->d_lease_hit);
299 
300 	return ret;
301 }
302 
303 void ceph_metric_destroy(struct ceph_client_metric *m)
304 {
305 	if (!m)
306 		return;
307 
308 	cancel_delayed_work_sync(&m->delayed_work);
309 
310 	percpu_counter_destroy(&m->total_inodes);
311 	percpu_counter_destroy(&m->opened_inodes);
312 	percpu_counter_destroy(&m->i_caps_mis);
313 	percpu_counter_destroy(&m->i_caps_hit);
314 	percpu_counter_destroy(&m->d_lease_mis);
315 	percpu_counter_destroy(&m->d_lease_hit);
316 
317 	ceph_put_mds_session(m->session);
318 }
319 
320 #define METRIC_UPDATE_MIN_MAX(min, max, new)	\
321 {						\
322 	if (unlikely(new < min))		\
323 		min = new;			\
324 	if (unlikely(new > max))		\
325 		max = new;			\
326 }
327 
328 static inline void __update_mean_and_stdev(ktime_t total, ktime_t *lavg,
329 					   ktime_t *sq_sump, ktime_t lat)
330 {
331 	ktime_t avg;
332 
333 	if (unlikely(total == 1)) {
334 		*lavg = lat;
335 	} else {
336 		/* the sq is (lat - old_avg) * (lat - new_avg) */
337 		avg = *lavg + div64_s64(lat - *lavg, total);
338 		*sq_sump += (lat - *lavg)*(lat - avg);
339 		*lavg = avg;
340 	}
341 }
342 
343 void ceph_update_metrics(struct ceph_metric *m,
344 			 ktime_t r_start, ktime_t r_end,
345 			 unsigned int size, int rc)
346 {
347 	ktime_t lat = ktime_sub(r_end, r_start);
348 	ktime_t total;
349 
350 	if (unlikely(rc < 0 && rc != -ENOENT && rc != -ETIMEDOUT))
351 		return;
352 
353 	spin_lock(&m->lock);
354 	total = ++m->total;
355 	m->size_sum += size;
356 	METRIC_UPDATE_MIN_MAX(m->size_min, m->size_max, size);
357 	m->latency_sum += lat;
358 	METRIC_UPDATE_MIN_MAX(m->latency_min, m->latency_max, lat);
359 	__update_mean_and_stdev(total, &m->latency_avg,	&m->latency_sq_sum,
360 				lat);
361 	spin_unlock(&m->lock);
362 }
363