1 |
tdb |
1.1 |
/* alloc_pool.c - routines for allocating memory from tagged pools */ |
2 |
|
|
|
3 |
|
|
/* Copyright 1991 Mark Russell, University of Kent at Canterbury. |
4 |
|
|
* |
5 |
|
|
* You can do what you like with this source code as long as |
6 |
|
|
* you don't try to make money out of it and you include an |
7 |
|
|
* unaltered copy of this message (including the copyright). |
8 |
|
|
*/ |
9 |
|
|
|
10 |
|
|
char ukcprog_alloc_sccsid[] = "$Id: alloc.c,v 1.10 1993/10/25 11:44:23 mtr Exp $ UKC"; |
11 |
|
|
|
12 |
|
|
#include <stdio.h> /* for NULL - grrr */ |
13 |
|
|
#include <stdlib.h> |
14 |
|
|
#include <string.h> |
15 |
|
|
#ifndef __STDC__ |
16 |
|
|
#include <memory.h> |
17 |
|
|
#endif |
18 |
|
|
|
19 |
|
|
#include "ukcprog.h" |
20 |
|
|
|
21 |
|
|
/* This is a conservative guess at the per-request malloc overhead in |
22 |
|
|
* bytes. Nothing breaks if this is wrong. |
23 |
|
|
*/ |
24 |
|
|
#define MALLOC_OVERHEAD 24 |
25 |
|
|
|
26 |
|
|
/* When we run out a space in an alloc pool we add another block. |
27 |
|
|
* We add small blocks (SBLOCKSIZE bytes each) for the first NSMALLBLOCKS |
28 |
|
|
* requests, then switch switch to large (BLOCKSIZE) ones. |
29 |
|
|
* |
30 |
|
|
* The intention is that we don't gobble large amounts of memory for |
31 |
|
|
* a small alloc pool, but that we are reasonablty efficient for |
32 |
|
|
* one that's continually growing. |
33 |
|
|
* |
34 |
|
|
* Currently, we go slowly (256 bytes a go) for the first 8K, then |
35 |
|
|
* fast (4K a go). |
36 |
|
|
*/ |
37 |
|
|
#define NSMALLBLOCKS 32 |
38 |
|
|
|
39 |
|
|
/* Size of the first block for an alloc pool (requested when the alloc |
40 |
|
|
* pool is created) and for the subsequent NSMALLBLOCKS blocks. |
41 |
|
|
*/ |
42 |
|
|
#define SBLOCKSIZE (256 - sizeof(alloc_pool_t) - MALLOC_OVERHEAD) |
43 |
|
|
|
44 |
|
|
/* Size of the requested for an alloc pool after the first NSMALLBLOCKS |
45 |
|
|
* block additions. |
46 |
|
|
* |
47 |
|
|
* Try to make the malloc request size a bit less than a power of two |
48 |
|
|
* to compensate for brain-damaged mallocs that add overhead then round |
49 |
|
|
* up to a power of two. |
50 |
|
|
*/ |
51 |
|
|
#define BLOCKSIZE (4096 - sizeof(ablock_t) - MALLOC_OVERHEAD) |
52 |
|
|
|
53 |
|
|
/* Maximum alignment requirements for all types *including* float and double. |
54 |
|
|
*/ |
55 |
|
|
#define ALIGN sizeof(double) |
56 |
|
|
|
57 |
|
|
typedef struct ablockst { |
58 |
|
|
union { |
59 |
|
|
double align; |
60 |
|
|
struct ablock { |
61 |
|
|
char *abu_buf; |
62 |
|
|
char *abu_pos; |
63 |
|
|
char *abu_end; |
64 |
|
|
size_t abu_size; |
65 |
|
|
struct ablockst *abu_next; |
66 |
|
|
} a; |
67 |
|
|
} u; |
68 |
|
|
} ablock_t; |
69 |
|
|
|
70 |
|
|
#define ab_buf u.a.abu_buf |
71 |
|
|
#define ab_pos u.a.abu_pos |
72 |
|
|
#define ab_end u.a.abu_end |
73 |
|
|
#define ab_size u.a.abu_size |
74 |
|
|
#define ab_next u.a.abu_next |
75 |
|
|
|
76 |
|
|
struct alloc_pool_s { |
77 |
|
|
ablock_t *ap_ablock; |
78 |
|
|
ablock_t *ap_freelist; |
79 |
|
|
int ap_nblocks; |
80 |
|
|
bool ap_debug; |
81 |
|
|
ablock_t ap_first_ablock; |
82 |
|
|
}; |
83 |
|
|
|
84 |
|
|
struct alloc_mark_s { |
85 |
|
|
alloc_pool_t *am_apool; |
86 |
|
|
ablock_t *am_ablock; |
87 |
|
|
char *am_pos; |
88 |
|
|
char *am_end; |
89 |
|
|
}; |
90 |
|
|
|
91 |
|
|
static ablock_t *push_ablock PROTO((alloc_pool_t *ap, ablock_t *ab, unsigned size)); |
92 |
|
|
static ablock_t *find_ab PROTO((alloc_pool_t *ap, unsigned size)); |
93 |
|
|
static void reset_ablocks PROTO((alloc_pool_t *ap, ablock_t *limab)); |
94 |
|
|
|
95 |
|
|
/* The default debug flag for a new alloc_pool. When the debug flag |
96 |
|
|
* is TRUE, we initialise memory to garbage, and set it to (different) |
97 |
|
|
* garbage when free_alloc_pool is called. |
98 |
|
|
*/ |
99 |
|
|
static bool Default_debug_flag = TRUE; |
100 |
|
|
|
101 |
|
|
bool |
102 |
|
|
alloc_set_default_debug_flag(val) |
103 |
|
|
bool val; |
104 |
|
|
{ |
105 |
|
|
bool oldval; |
106 |
|
|
|
107 |
|
|
oldval = Default_debug_flag; |
108 |
|
|
Default_debug_flag = val; |
109 |
|
|
return oldval; |
110 |
|
|
} |
111 |
|
|
|
112 |
|
|
bool |
113 |
|
|
alloc_set_debug_flag(ap, val) |
114 |
|
|
alloc_pool_t *ap; |
115 |
|
|
bool val; |
116 |
|
|
{ |
117 |
|
|
bool oldval; |
118 |
|
|
|
119 |
|
|
oldval = ap->ap_debug; |
120 |
|
|
ap->ap_debug = val; |
121 |
|
|
return oldval; |
122 |
|
|
} |
123 |
|
|
|
124 |
|
|
/* Make a new alloc_pool(). We make an initial allocation of a small |
125 |
|
|
* amount of memory, to make small alloc pool creation cheap (one malloc). |
126 |
|
|
*/ |
127 |
|
|
alloc_pool_t * |
128 |
|
|
alloc_create_pool() |
129 |
|
|
{ |
130 |
|
|
alloc_pool_t *ap; |
131 |
|
|
|
132 |
|
|
ap = (alloc_pool_t *)e_malloc(sizeof(alloc_pool_t) + SBLOCKSIZE); |
133 |
|
|
ap->ap_ablock = NULL; |
134 |
|
|
ap->ap_freelist = NULL; |
135 |
|
|
ap->ap_nblocks = 0; |
136 |
|
|
ap->ap_debug = Default_debug_flag; |
137 |
|
|
push_ablock(ap, &ap->ap_first_ablock, SBLOCKSIZE); |
138 |
|
|
|
139 |
|
|
return ap; |
140 |
|
|
} |
141 |
|
|
|
142 |
|
|
static void |
143 |
|
|
reset_ablocks(ap, limab) |
144 |
|
|
alloc_pool_t *ap; |
145 |
|
|
ablock_t *limab; |
146 |
|
|
{ |
147 |
|
|
ablock_t *ab, *next; |
148 |
|
|
bool debug; |
149 |
|
|
|
150 |
|
|
debug = ap->ap_debug; |
151 |
|
|
for (ab = ap->ap_ablock; ab != limab; ab = next) { |
152 |
|
|
next = ab->ab_next; |
153 |
|
|
if (debug) |
154 |
|
|
memset(ab->ab_buf, 0x42, ab->ab_size); |
155 |
|
|
ab->ab_pos = ab->ab_buf; |
156 |
|
|
ab->ab_end = ab->ab_pos + ab->ab_size; |
157 |
|
|
ab->ab_next = ap->ap_freelist; |
158 |
|
|
ap->ap_freelist = ab; |
159 |
|
|
} |
160 |
|
|
} |
161 |
|
|
|
162 |
|
|
void |
163 |
|
|
alloc_reset_pool(ap) |
164 |
|
|
alloc_pool_t *ap; |
165 |
|
|
{ |
166 |
|
|
ablock_t *ab; |
167 |
|
|
|
168 |
|
|
ab = &ap->ap_first_ablock; |
169 |
|
|
|
170 |
|
|
reset_ablocks(ap, ab); |
171 |
|
|
|
172 |
|
|
if (ap->ap_debug) |
173 |
|
|
memset(ab->ab_buf, 0x42, ab->ab_size); |
174 |
|
|
ab->ab_pos = ab->ab_buf; |
175 |
|
|
ab->ab_end = ab->ab_pos + ab->ab_size; |
176 |
|
|
|
177 |
|
|
ap->ap_ablock = ab; |
178 |
|
|
} |
179 |
|
|
|
180 |
|
|
void |
181 |
|
|
alloc_free_pool(ap) |
182 |
|
|
alloc_pool_t *ap; |
183 |
|
|
{ |
184 |
|
|
ablock_t *ab, *next; |
185 |
|
|
bool debug; |
186 |
|
|
|
187 |
|
|
debug = ap->ap_debug; |
188 |
|
|
|
189 |
|
|
/* The odd form of the loop here is because we want to overwrite |
190 |
|
|
* all blocks with garbage (if debug is set), but we don't want |
191 |
|
|
* to free the last block in the chain, which is allocated as part |
192 |
|
|
* of the header block. |
193 |
|
|
*/ |
194 |
|
|
ab = ap->ap_ablock; |
195 |
|
|
for (;;) { |
196 |
|
|
next = ab->ab_next; |
197 |
|
|
if (debug) |
198 |
|
|
memset(ab->ab_buf, 0x42, ab->ab_size); |
199 |
|
|
if (next == NULL) |
200 |
|
|
break; |
201 |
|
|
free((char *)ab); |
202 |
|
|
ab = next; |
203 |
|
|
} |
204 |
|
|
|
205 |
|
|
free((char *)ap); |
206 |
|
|
} |
207 |
|
|
|
208 |
|
|
static ablock_t * |
209 |
|
|
push_ablock(ap, ab, size) |
210 |
|
|
alloc_pool_t *ap; |
211 |
|
|
ablock_t *ab; |
212 |
|
|
unsigned size; |
213 |
|
|
{ |
214 |
|
|
ab->ab_buf = ab->ab_pos = (char *)&ab[1]; |
215 |
|
|
ab->ab_end = ab->ab_buf + size; |
216 |
|
|
ab->ab_size = size; |
217 |
|
|
ab->ab_next = ap->ap_ablock; |
218 |
|
|
ap->ap_ablock = ab; |
219 |
|
|
|
220 |
|
|
if (ap->ap_debug) |
221 |
|
|
memset(ab->ab_buf, 0x53, (size_t)size); |
222 |
|
|
|
223 |
|
|
return ab; |
224 |
|
|
} |
225 |
|
|
|
226 |
|
|
/* Find an ablock with at least nbytes free. If the block at the |
227 |
|
|
* head of the free list is big enough, use that. Otherwise malloc |
228 |
|
|
* a new ablock and push it on the chain. |
229 |
|
|
*/ |
230 |
|
|
static ablock_t * |
231 |
|
|
find_ab(ap, size) |
232 |
|
|
alloc_pool_t *ap; |
233 |
|
|
unsigned size; |
234 |
|
|
{ |
235 |
|
|
ablock_t *ab; |
236 |
|
|
|
237 |
|
|
if (ap->ap_freelist != NULL && ap->ap_freelist->ab_size >= size) { |
238 |
|
|
ab = ap->ap_freelist; |
239 |
|
|
ap->ap_freelist = ap->ap_freelist->ab_next; |
240 |
|
|
ab->ab_next = ap->ap_ablock; |
241 |
|
|
ap->ap_ablock = ab; |
242 |
|
|
} |
243 |
|
|
else { |
244 |
|
|
voidptr buf; |
245 |
|
|
unsigned blocksize; |
246 |
|
|
|
247 |
|
|
blocksize = (ap->ap_nblocks < NSMALLBLOCKS) ? SBLOCKSIZE : BLOCKSIZE; |
248 |
|
|
if (size < blocksize) |
249 |
|
|
size = blocksize; |
250 |
|
|
if ((buf = malloc((size_t)(sizeof(ablock_t) + size))) == NULL) |
251 |
|
|
return NULL; |
252 |
|
|
ab = push_ablock(ap, (ablock_t *)buf, size); |
253 |
|
|
++ap->ap_nblocks; |
254 |
|
|
} |
255 |
|
|
return ab; |
256 |
|
|
} |
257 |
|
|
|
258 |
|
|
/* Allocate nbytes from alloc pool ap. This interface never |
259 |
|
|
* returns NULL - if memory runs out we panic. |
260 |
|
|
*/ |
261 |
|
|
voidptr |
262 |
|
|
alloc(ap, nbytes) |
263 |
|
|
alloc_pool_t *ap; |
264 |
|
|
size_t nbytes; |
265 |
|
|
{ |
266 |
|
|
ablock_t *ab; |
267 |
|
|
int over; |
268 |
|
|
char *ptr; |
269 |
|
|
|
270 |
|
|
over = nbytes % ALIGN; |
271 |
|
|
if (over != 0) |
272 |
|
|
nbytes += ALIGN - over; |
273 |
|
|
|
274 |
|
|
ab = ap->ap_ablock; |
275 |
|
|
|
276 |
|
|
if (nbytes > ab->ab_end - ab->ab_pos) { |
277 |
|
|
ab = find_ab(ap, (unsigned)nbytes); |
278 |
|
|
if (ab == NULL) |
279 |
|
|
panic("out of memory in alloc"); |
280 |
|
|
} |
281 |
|
|
|
282 |
|
|
ptr = ab->ab_pos; |
283 |
|
|
ab->ab_pos += nbytes; |
284 |
|
|
|
285 |
|
|
return ptr; |
286 |
|
|
} |
287 |
|
|
|
288 |
|
|
/* Like alloc, but return NULL if we can't satisfy the request. |
289 |
|
|
*/ |
290 |
|
|
voidptr |
291 |
|
|
alloc_ck(ap, nbytes) |
292 |
|
|
alloc_pool_t *ap; |
293 |
|
|
size_t nbytes; |
294 |
|
|
{ |
295 |
|
|
ablock_t *ab; |
296 |
|
|
int over; |
297 |
|
|
char *ptr; |
298 |
|
|
|
299 |
|
|
over = nbytes % ALIGN; |
300 |
|
|
if (over != 0) |
301 |
|
|
nbytes += ALIGN - over; |
302 |
|
|
|
303 |
|
|
ab = ap->ap_ablock; |
304 |
|
|
|
305 |
|
|
if (nbytes > ab->ab_end - ab->ab_pos) { |
306 |
|
|
ab = find_ab(ap, nbytes); |
307 |
|
|
if (ab == NULL) |
308 |
|
|
return NULL; |
309 |
|
|
} |
310 |
|
|
|
311 |
|
|
ptr = ab->ab_pos; |
312 |
|
|
ab->ab_pos += nbytes; |
313 |
|
|
|
314 |
|
|
return ptr; |
315 |
|
|
} |
316 |
|
|
|
317 |
|
|
alloc_mark_t * |
318 |
|
|
alloc_mark(ap) |
319 |
|
|
alloc_pool_t *ap; |
320 |
|
|
{ |
321 |
|
|
alloc_mark_t *am; |
322 |
|
|
ablock_t *save_ab; |
323 |
|
|
char *save_pos, *save_end; |
324 |
|
|
|
325 |
|
|
save_ab = ap->ap_ablock; |
326 |
|
|
save_pos = save_ab->ab_pos; |
327 |
|
|
save_end = save_ab->ab_end; |
328 |
|
|
|
329 |
|
|
am = (alloc_mark_t *)alloc(ap, sizeof(alloc_mark_t)); |
330 |
|
|
am->am_apool = ap; |
331 |
|
|
am->am_ablock = save_ab; |
332 |
|
|
am->am_pos = save_pos; |
333 |
|
|
am->am_end = save_end; |
334 |
|
|
|
335 |
|
|
return am; |
336 |
|
|
} |
337 |
|
|
|
338 |
|
|
void |
339 |
|
|
alloc_release(ap, am) |
340 |
|
|
alloc_pool_t *ap; |
341 |
|
|
alloc_mark_t *am; |
342 |
|
|
{ |
343 |
|
|
ablock_t *ab; |
344 |
|
|
alloc_mark_t mark; |
345 |
|
|
|
346 |
|
|
if (am->am_apool != ap) |
347 |
|
|
panic("id botch in ar"); |
348 |
|
|
|
349 |
|
|
/* If debug is set, we are about to step on the store that |
350 |
|
|
* the mark was allocated from, so save it. |
351 |
|
|
*/ |
352 |
|
|
mark = *am; |
353 |
|
|
ab = mark.am_ablock; |
354 |
|
|
|
355 |
|
|
reset_ablocks(ap, ab); |
356 |
|
|
|
357 |
|
|
if (ap->ap_debug) { |
358 |
|
|
memset(mark.am_pos, 0x42, (size_t)(ab->ab_pos - mark.am_pos)); |
359 |
|
|
memset(ab->ab_end, 0x42, (size_t)(mark.am_end - ab->ab_end)); |
360 |
|
|
} |
361 |
|
|
else { |
362 |
|
|
/* Make sure the application can't use this mark again. |
363 |
|
|
*/ |
364 |
|
|
am->am_apool = NULL; |
365 |
|
|
} |
366 |
|
|
|
367 |
|
|
ab->ab_pos = mark.am_pos; |
368 |
|
|
ab->ab_end = mark.am_end; |
369 |
|
|
ap->ap_ablock = ab; |
370 |
|
|
} |
371 |
|
|
|
372 |
|
|
/* Like alloc(), except that the result is assumed not to need alignment. |
373 |
|
|
* We work from the other end of the pool than alloc so hopefully all the |
374 |
|
|
* string requests will be packed together with no alignment holes. |
375 |
|
|
* |
376 |
|
|
* We never return NULL - if we can't fulfill the request we panic. |
377 |
|
|
*/ |
378 |
|
|
char * |
379 |
|
|
allocstr(ap, nbytes) |
380 |
|
|
alloc_pool_t *ap; |
381 |
|
|
size_t nbytes; |
382 |
|
|
{ |
383 |
|
|
ablock_t *ab; |
384 |
|
|
|
385 |
|
|
ab = ap->ap_ablock; |
386 |
|
|
|
387 |
|
|
if (nbytes > ab->ab_end - ab->ab_pos) { |
388 |
|
|
ab = find_ab(ap, (unsigned)nbytes); |
389 |
|
|
if (ab == NULL) |
390 |
|
|
panic("out of memory in allocstr"); |
391 |
|
|
} |
392 |
|
|
|
393 |
|
|
return ab->ab_end -= nbytes; |
394 |
|
|
} |
395 |
|
|
|
396 |
|
|
char * |
397 |
|
|
allocstr_ck(ap, nbytes) |
398 |
|
|
alloc_pool_t *ap; |
399 |
|
|
size_t nbytes; |
400 |
|
|
{ |
401 |
|
|
ablock_t *ab; |
402 |
|
|
|
403 |
|
|
ab = ap->ap_ablock; |
404 |
|
|
|
405 |
|
|
/* We cast nbytes to unsigned to catch negative values: they |
406 |
|
|
* turn into huge positive values which get caught by e_malloc(). |
407 |
|
|
*/ |
408 |
|
|
if ((unsigned)nbytes > ab->ab_end - ab->ab_pos) { |
409 |
|
|
ab = find_ab(ap, (unsigned)nbytes); |
410 |
|
|
if (ab == NULL) |
411 |
|
|
return NULL; |
412 |
|
|
} |
413 |
|
|
|
414 |
|
|
return ab->ab_end -= nbytes; |
415 |
|
|
} |
416 |
|
|
|
417 |
|
|
char * |
418 |
|
|
alloc_strdup(ap, s) |
419 |
|
|
alloc_pool_t *ap; |
420 |
|
|
const char *s; |
421 |
|
|
{ |
422 |
|
|
size_t nbytes; |
423 |
|
|
|
424 |
|
|
nbytes = strlen(s) + 1; |
425 |
|
|
return memcpy(allocstr(ap, nbytes), s, nbytes); |
426 |
|
|
} |