xref: /linux/lib/lzo/lzo1x_compress.c (revision 0da908c291070d89482f6211dbe81d4d43c3f7cb)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  *  LZO1X Compressor from LZO
4  *
5  *  Copyright (C) 1996-2012 Markus F.X.J. Oberhumer <markus@oberhumer.com>
6  *
7  *  The full LZO package can be found at:
8  *  http://www.oberhumer.com/opensource/lzo/
9  *
10  *  Changed for Linux kernel use by:
11  *  Nitin Gupta <nitingupta910@gmail.com>
12  *  Richard Purdie <rpurdie@openedhand.com>
13  */
14 
15 #include <linux/module.h>
16 #include <linux/kernel.h>
17 #include <asm/unaligned.h>
18 #include <linux/lzo.h>
19 #include "lzodefs.h"
20 
21 static noinline size_t
22 lzo1x_1_do_compress(const unsigned char *in, size_t in_len,
23 		    unsigned char *out, size_t *out_len,
24 		    size_t ti, void *wrkmem, signed char *state_offset,
25 		    const unsigned char bitstream_version)
26 {
27 	const unsigned char *ip;
28 	unsigned char *op;
29 	const unsigned char * const in_end = in + in_len;
30 	const unsigned char * const ip_end = in + in_len - 20;
31 	const unsigned char *ii;
32 	lzo_dict_t * const dict = (lzo_dict_t *) wrkmem;
33 
34 	op = out;
35 	ip = in;
36 	ii = ip;
37 	ip += ti < 4 ? 4 - ti : 0;
38 
39 	for (;;) {
40 		const unsigned char *m_pos = NULL;
41 		size_t t, m_len, m_off;
42 		u32 dv;
43 		u32 run_length = 0;
44 literal:
45 		ip += 1 + ((ip - ii) >> 5);
46 next:
47 		if (unlikely(ip >= ip_end))
48 			break;
49 		dv = get_unaligned_le32(ip);
50 
51 		if (dv == 0 && bitstream_version) {
52 			const unsigned char *ir = ip + 4;
53 			const unsigned char *limit = min(ip_end, ip + MAX_ZERO_RUN_LENGTH + 1);
54 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && \
55 	defined(LZO_FAST_64BIT_MEMORY_ACCESS)
56 			u64 dv64;
57 
58 			for (; (ir + 32) <= limit; ir += 32) {
59 				dv64 = get_unaligned((u64 *)ir);
60 				dv64 |= get_unaligned((u64 *)ir + 1);
61 				dv64 |= get_unaligned((u64 *)ir + 2);
62 				dv64 |= get_unaligned((u64 *)ir + 3);
63 				if (dv64)
64 					break;
65 			}
66 			for (; (ir + 8) <= limit; ir += 8) {
67 				dv64 = get_unaligned((u64 *)ir);
68 				if (dv64) {
69 #  if defined(__LITTLE_ENDIAN)
70 					ir += __builtin_ctzll(dv64) >> 3;
71 #  elif defined(__BIG_ENDIAN)
72 					ir += __builtin_clzll(dv64) >> 3;
73 #  else
74 #    error "missing endian definition"
75 #  endif
76 					break;
77 				}
78 			}
79 #else
80 			while ((ir < (const unsigned char *)
81 					ALIGN((uintptr_t)ir, 4)) &&
82 					(ir < limit) && (*ir == 0))
83 				ir++;
84 			if (IS_ALIGNED((uintptr_t)ir, 4)) {
85 				for (; (ir + 4) <= limit; ir += 4) {
86 					dv = *((u32 *)ir);
87 					if (dv) {
88 #  if defined(__LITTLE_ENDIAN)
89 						ir += __builtin_ctz(dv) >> 3;
90 #  elif defined(__BIG_ENDIAN)
91 						ir += __builtin_clz(dv) >> 3;
92 #  else
93 #    error "missing endian definition"
94 #  endif
95 						break;
96 					}
97 				}
98 			}
99 #endif
100 			while (likely(ir < limit) && unlikely(*ir == 0))
101 				ir++;
102 			run_length = ir - ip;
103 			if (run_length > MAX_ZERO_RUN_LENGTH)
104 				run_length = MAX_ZERO_RUN_LENGTH;
105 		} else {
106 			t = ((dv * 0x1824429d) >> (32 - D_BITS)) & D_MASK;
107 			m_pos = in + dict[t];
108 			dict[t] = (lzo_dict_t) (ip - in);
109 			if (unlikely(dv != get_unaligned_le32(m_pos)))
110 				goto literal;
111 		}
112 
113 		ii -= ti;
114 		ti = 0;
115 		t = ip - ii;
116 		if (t != 0) {
117 			if (t <= 3) {
118 				op[*state_offset] |= t;
119 				COPY4(op, ii);
120 				op += t;
121 			} else if (t <= 16) {
122 				*op++ = (t - 3);
123 				COPY8(op, ii);
124 				COPY8(op + 8, ii + 8);
125 				op += t;
126 			} else {
127 				if (t <= 18) {
128 					*op++ = (t - 3);
129 				} else {
130 					size_t tt = t - 18;
131 					*op++ = 0;
132 					while (unlikely(tt > 255)) {
133 						tt -= 255;
134 						*op++ = 0;
135 					}
136 					*op++ = tt;
137 				}
138 				do {
139 					COPY8(op, ii);
140 					COPY8(op + 8, ii + 8);
141 					op += 16;
142 					ii += 16;
143 					t -= 16;
144 				} while (t >= 16);
145 				if (t > 0) do {
146 					*op++ = *ii++;
147 				} while (--t > 0);
148 			}
149 		}
150 
151 		if (unlikely(run_length)) {
152 			ip += run_length;
153 			run_length -= MIN_ZERO_RUN_LENGTH;
154 			put_unaligned_le32((run_length << 21) | 0xfffc18
155 					   | (run_length & 0x7), op);
156 			op += 4;
157 			run_length = 0;
158 			*state_offset = -3;
159 			goto finished_writing_instruction;
160 		}
161 
162 		m_len = 4;
163 		{
164 #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && defined(LZO_USE_CTZ64)
165 		u64 v;
166 		v = get_unaligned((const u64 *) (ip + m_len)) ^
167 		    get_unaligned((const u64 *) (m_pos + m_len));
168 		if (unlikely(v == 0)) {
169 			do {
170 				m_len += 8;
171 				v = get_unaligned((const u64 *) (ip + m_len)) ^
172 				    get_unaligned((const u64 *) (m_pos + m_len));
173 				if (unlikely(ip + m_len >= ip_end))
174 					goto m_len_done;
175 			} while (v == 0);
176 		}
177 #  if defined(__LITTLE_ENDIAN)
178 		m_len += (unsigned) __builtin_ctzll(v) / 8;
179 #  elif defined(__BIG_ENDIAN)
180 		m_len += (unsigned) __builtin_clzll(v) / 8;
181 #  else
182 #    error "missing endian definition"
183 #  endif
184 #elif defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && defined(LZO_USE_CTZ32)
185 		u32 v;
186 		v = get_unaligned((const u32 *) (ip + m_len)) ^
187 		    get_unaligned((const u32 *) (m_pos + m_len));
188 		if (unlikely(v == 0)) {
189 			do {
190 				m_len += 4;
191 				v = get_unaligned((const u32 *) (ip + m_len)) ^
192 				    get_unaligned((const u32 *) (m_pos + m_len));
193 				if (v != 0)
194 					break;
195 				m_len += 4;
196 				v = get_unaligned((const u32 *) (ip + m_len)) ^
197 				    get_unaligned((const u32 *) (m_pos + m_len));
198 				if (unlikely(ip + m_len >= ip_end))
199 					goto m_len_done;
200 			} while (v == 0);
201 		}
202 #  if defined(__LITTLE_ENDIAN)
203 		m_len += (unsigned) __builtin_ctz(v) / 8;
204 #  elif defined(__BIG_ENDIAN)
205 		m_len += (unsigned) __builtin_clz(v) / 8;
206 #  else
207 #    error "missing endian definition"
208 #  endif
209 #else
210 		if (unlikely(ip[m_len] == m_pos[m_len])) {
211 			do {
212 				m_len += 1;
213 				if (ip[m_len] != m_pos[m_len])
214 					break;
215 				m_len += 1;
216 				if (ip[m_len] != m_pos[m_len])
217 					break;
218 				m_len += 1;
219 				if (ip[m_len] != m_pos[m_len])
220 					break;
221 				m_len += 1;
222 				if (ip[m_len] != m_pos[m_len])
223 					break;
224 				m_len += 1;
225 				if (ip[m_len] != m_pos[m_len])
226 					break;
227 				m_len += 1;
228 				if (ip[m_len] != m_pos[m_len])
229 					break;
230 				m_len += 1;
231 				if (ip[m_len] != m_pos[m_len])
232 					break;
233 				m_len += 1;
234 				if (unlikely(ip + m_len >= ip_end))
235 					goto m_len_done;
236 			} while (ip[m_len] == m_pos[m_len]);
237 		}
238 #endif
239 		}
240 m_len_done:
241 
242 		m_off = ip - m_pos;
243 		ip += m_len;
244 		if (m_len <= M2_MAX_LEN && m_off <= M2_MAX_OFFSET) {
245 			m_off -= 1;
246 			*op++ = (((m_len - 1) << 5) | ((m_off & 7) << 2));
247 			*op++ = (m_off >> 3);
248 		} else if (m_off <= M3_MAX_OFFSET) {
249 			m_off -= 1;
250 			if (m_len <= M3_MAX_LEN)
251 				*op++ = (M3_MARKER | (m_len - 2));
252 			else {
253 				m_len -= M3_MAX_LEN;
254 				*op++ = M3_MARKER | 0;
255 				while (unlikely(m_len > 255)) {
256 					m_len -= 255;
257 					*op++ = 0;
258 				}
259 				*op++ = (m_len);
260 			}
261 			*op++ = (m_off << 2);
262 			*op++ = (m_off >> 6);
263 		} else {
264 			m_off -= 0x4000;
265 			if (m_len <= M4_MAX_LEN)
266 				*op++ = (M4_MARKER | ((m_off >> 11) & 8)
267 						| (m_len - 2));
268 			else {
269 				if (unlikely(((m_off & 0x403f) == 0x403f)
270 						&& (m_len >= 261)
271 						&& (m_len <= 264))
272 						&& likely(bitstream_version)) {
273 					// Under lzo-rle, block copies
274 					// for 261 <= length <= 264 and
275 					// (distance & 0x80f3) == 0x80f3
276 					// can result in ambiguous
277 					// output. Adjust length
278 					// to 260 to prevent ambiguity.
279 					ip -= m_len - 260;
280 					m_len = 260;
281 				}
282 				m_len -= M4_MAX_LEN;
283 				*op++ = (M4_MARKER | ((m_off >> 11) & 8));
284 				while (unlikely(m_len > 255)) {
285 					m_len -= 255;
286 					*op++ = 0;
287 				}
288 				*op++ = (m_len);
289 			}
290 			*op++ = (m_off << 2);
291 			*op++ = (m_off >> 6);
292 		}
293 		*state_offset = -2;
294 finished_writing_instruction:
295 		ii = ip;
296 		goto next;
297 	}
298 	*out_len = op - out;
299 	return in_end - (ii - ti);
300 }
301 
302 static int lzogeneric1x_1_compress(const unsigned char *in, size_t in_len,
303 		     unsigned char *out, size_t *out_len,
304 		     void *wrkmem, const unsigned char bitstream_version)
305 {
306 	const unsigned char *ip = in;
307 	unsigned char *op = out;
308 	unsigned char *data_start;
309 	size_t l = in_len;
310 	size_t t = 0;
311 	signed char state_offset = -2;
312 	unsigned int m4_max_offset;
313 
314 	// LZO v0 will never write 17 as first byte (except for zero-length
315 	// input), so this is used to version the bitstream
316 	if (bitstream_version > 0) {
317 		*op++ = 17;
318 		*op++ = bitstream_version;
319 		m4_max_offset = M4_MAX_OFFSET_V1;
320 	} else {
321 		m4_max_offset = M4_MAX_OFFSET_V0;
322 	}
323 
324 	data_start = op;
325 
326 	while (l > 20) {
327 		size_t ll = min_t(size_t, l, m4_max_offset + 1);
328 		uintptr_t ll_end = (uintptr_t) ip + ll;
329 		if ((ll_end + ((t + ll) >> 5)) <= ll_end)
330 			break;
331 		BUILD_BUG_ON(D_SIZE * sizeof(lzo_dict_t) > LZO1X_1_MEM_COMPRESS);
332 		memset(wrkmem, 0, D_SIZE * sizeof(lzo_dict_t));
333 		t = lzo1x_1_do_compress(ip, ll, op, out_len, t, wrkmem,
334 					&state_offset, bitstream_version);
335 		ip += ll;
336 		op += *out_len;
337 		l  -= ll;
338 	}
339 	t += l;
340 
341 	if (t > 0) {
342 		const unsigned char *ii = in + in_len - t;
343 
344 		if (op == data_start && t <= 238) {
345 			*op++ = (17 + t);
346 		} else if (t <= 3) {
347 			op[state_offset] |= t;
348 		} else if (t <= 18) {
349 			*op++ = (t - 3);
350 		} else {
351 			size_t tt = t - 18;
352 			*op++ = 0;
353 			while (tt > 255) {
354 				tt -= 255;
355 				*op++ = 0;
356 			}
357 			*op++ = tt;
358 		}
359 		if (t >= 16) do {
360 			COPY8(op, ii);
361 			COPY8(op + 8, ii + 8);
362 			op += 16;
363 			ii += 16;
364 			t -= 16;
365 		} while (t >= 16);
366 		if (t > 0) do {
367 			*op++ = *ii++;
368 		} while (--t > 0);
369 	}
370 
371 	*op++ = M4_MARKER | 1;
372 	*op++ = 0;
373 	*op++ = 0;
374 
375 	*out_len = op - out;
376 	return LZO_E_OK;
377 }
378 
379 int lzo1x_1_compress(const unsigned char *in, size_t in_len,
380 		     unsigned char *out, size_t *out_len,
381 		     void *wrkmem)
382 {
383 	return lzogeneric1x_1_compress(in, in_len, out, out_len, wrkmem, 0);
384 }
385 
386 int lzorle1x_1_compress(const unsigned char *in, size_t in_len,
387 		     unsigned char *out, size_t *out_len,
388 		     void *wrkmem)
389 {
390 	return lzogeneric1x_1_compress(in, in_len, out, out_len,
391 				       wrkmem, LZO_VERSION);
392 }
393 
394 EXPORT_SYMBOL_GPL(lzo1x_1_compress);
395 EXPORT_SYMBOL_GPL(lzorle1x_1_compress);
396 
397 MODULE_LICENSE("GPL");
398 MODULE_DESCRIPTION("LZO1X-1 Compressor");
399