-
Notifications
You must be signed in to change notification settings - Fork 560
/
Copy pathmalloc.c
2407 lines (2150 loc) · 80.5 KB
/
malloc.c
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
/* malloc.c
*
*/
/*
* 'The Chamber of Records,' said Gimli. 'I guess that is where we now stand.'
*
* [p.321 of _The Lord of the Rings_, II/v: "The Bridge of Khazad-Dûm"]
*/
/* This file contains Perl's own implementation of the malloc library.
* It is used if Configure decides that, on your platform, Perl's
* version is better than the OS's, or if you give Configure the
* -Dusemymalloc command-line option.
*/
/*
Here are some notes on configuring Perl's malloc.
There are two macros which serve as bulk disablers of advanced
features of this malloc: NO_FANCY_MALLOC, PLAIN_MALLOC (undef by
default). Look in the list of default values below to understand
their exact effect. Defining NO_FANCY_MALLOC returns malloc.c to the
state of the malloc in Perl 5.004. Additionally defining PLAIN_MALLOC
returns it to the state as of Perl 5.000.
Note that some of the settings below may be ignored in the code based
on values of other macros. The PERL_CORE symbol is only defined when
perl itself is being compiled (so malloc can make some assumptions
about perl's facilities being available to it).
Each config option has a short description, followed by its name,
default value, and a comment about the default (if applicable). Some
options take a precise value, while the others are just boolean.
The boolean ones are listed first.
# Read configuration settings from malloc_cfg.h
HAVE_MALLOC_CFG_H undef
# Enable code for an emergency memory pool in $^M. See perlvar.pod
# for a description of $^M.
PERL_EMERGENCY_SBRK !PLAIN_MALLOC
# Enable code for printing memory statistics.
DEBUGGING_MSTATS !PLAIN_MALLOC
# Move allocation info for small buckets into separate areas.
# Memory optimization (especially for small allocations, of the
# less than 64 bytes). Since perl usually makes a large number
# of small allocations, this is usually a win.
PACK_MALLOC (!PLAIN_MALLOC && !RCHECK)
# Add one page to big powers of two when calculating bucket size.
# This is targeted at big allocations, as are common in image
# processing.
TWO_POT_OPTIMIZE !PLAIN_MALLOC
# Use intermediate bucket sizes between powers-of-two. This is
# generally a memory optimization, and a (small) speed pessimization.
BUCKETS_ROOT2 !NO_FANCY_MALLOC
# Do not check small deallocations for bad free(). Memory
# and speed optimization, error reporting pessimization.
IGNORE_SMALL_BAD_FREE (!NO_FANCY_MALLOC && !RCHECK)
# Use table lookup to decide in which bucket a given allocation will go.
SMALL_BUCKET_VIA_TABLE !NO_FANCY_MALLOC
# Use a perl-defined sbrk() instead of the (presumably broken or
# missing) system-supplied sbrk().
USE_PERL_SBRK undef
# Use system malloc() (or calloc() etc.) to emulate sbrk(). Normally
# only used with broken sbrk()s.
PERL_SBRK_VIA_MALLOC undef
# Which allocator to use if PERL_SBRK_VIA_MALLOC
SYSTEM_ALLOC(a) malloc(a)
# Minimal alignment (in bytes, should be a power of 2) of SYSTEM_ALLOC
SYSTEM_ALLOC_ALIGNMENT MEM_ALIGNBYTES
# Disable memory overwrite checking with DEBUGGING. Memory and speed
# optimization, error reporting pessimization.
NO_RCHECK undef
# Enable memory overwrite checking with DEBUGGING. Memory and speed
# pessimization, error reporting optimization
RCHECK (DEBUGGING && !NO_RCHECK)
# Do not overwrite uninit areas with DEBUGGING. Speed
# optimization, error reporting pessimization
NO_MFILL undef
# Overwrite uninit areas with DEBUGGING. Speed
# pessimization, error reporting optimization
MALLOC_FILL (DEBUGGING && !NO_RCHECK && !NO_MFILL)
# Do not check overwritten uninit areas with DEBUGGING. Speed
# optimization, error reporting pessimization
NO_FILL_CHECK undef
# Check overwritten uninit areas with DEBUGGING. Speed
# pessimization, error reporting optimization
MALLOC_FILL_CHECK (DEBUGGING && !NO_RCHECK && !NO_FILL_CHECK)
# Failed allocations bigger than this size croak (if
# PERL_EMERGENCY_SBRK is enabled) without touching $^M. See
# perlvar.pod for a description of $^M.
BIG_SIZE (1<<16) # 64K
# Starting from this power of two, add an extra page to the
# size of the bucket. This enables optimized allocations of sizes
# close to powers of 2. Note that the value is indexed at 0.
FIRST_BIG_POW2 15 # 32K, 16K is used too often
# Estimate of minimal memory footprint. malloc uses this value to
# request the most reasonable largest blocks of memory from the system.
FIRST_SBRK (48*1024)
# Round up sbrk()s to multiples of this.
MIN_SBRK 2048
# Round up sbrk()s to multiples of this percent of footprint.
MIN_SBRK_FRAC 3
# Round up sbrk()s to multiples of this multiple of 1/1000 of footprint.
MIN_SBRK_FRAC1000 (10 * MIN_SBRK_FRAC)
# Add this much memory to big powers of two to get the bucket size.
PERL_PAGESIZE 4096
# This many sbrk() discontinuities should be tolerated even
# from the start without deciding that sbrk() is usually
# discontinuous.
SBRK_ALLOW_FAILURES 3
# This many continuous sbrk()s compensate for one discontinuous one.
SBRK_FAILURE_PRICE 50
# Some configurations may ask for 12-byte-or-so allocations which
# require 8-byte alignment (?!). In such situation one needs to
# define this to disable 12-byte bucket (will increase memory footprint)
STRICT_ALIGNMENT undef
# Do not allow configuration of runtime options at runtime
NO_MALLOC_DYNAMIC_CFG undef
# Do not allow configuration of runtime options via $ENV{PERL_MALLOC_OPT}
NO_PERL_MALLOC_ENV undef
[The variable consists of ;-separated parts of the form CODE=VALUE
with 1-character codes F, M, f, A, P, G, d, a, c for runtime
configuration of FIRST_SBRK, MIN_SBRK, MIN_SBRK_FRAC1000,
SBRK_ALLOW_FAILURES, SBRK_FAILURE_PRICE, sbrk_goodness,
filldead, fillalive, fillcheck. The last 3 are for DEBUGGING
build, and allow switching the tests for free()ed memory read,
uninit memory reads, and free()ed memory write.]
This implementation assumes that calling PerlIO_printf() does not
result in any memory allocation calls (used during a panic).
*/
#ifdef HAVE_MALLOC_CFG_H
# include "malloc_cfg.h"
#endif
#ifndef NO_FANCY_MALLOC
# ifndef SMALL_BUCKET_VIA_TABLE
# define SMALL_BUCKET_VIA_TABLE
# endif
# ifndef BUCKETS_ROOT2
# define BUCKETS_ROOT2
# endif
# ifndef IGNORE_SMALL_BAD_FREE
# define IGNORE_SMALL_BAD_FREE
# endif
#endif
#ifndef PLAIN_MALLOC /* Bulk enable features */
# ifndef PACK_MALLOC
# define PACK_MALLOC
# endif
# ifndef TWO_POT_OPTIMIZE
# define TWO_POT_OPTIMIZE
# endif
# ifndef PERL_EMERGENCY_SBRK
# define PERL_EMERGENCY_SBRK
# endif
# ifndef DEBUGGING_MSTATS
# define DEBUGGING_MSTATS
# endif
#endif
#define MIN_BUC_POW2 (sizeof(void*) > 4 ? 3 : 2) /* Allow for 4-byte arena. */
#define MIN_BUCKET (MIN_BUC_POW2 * BUCKETS_PER_POW2)
#define LOG_OF_MIN_ARENA 11
#if defined(DEBUGGING) && !defined(NO_RCHECK)
# define RCHECK
#endif
#if defined(DEBUGGING) && !defined(NO_RCHECK) && !defined(NO_MFILL) && !defined(MALLOC_FILL)
# define MALLOC_FILL
#endif
#if defined(DEBUGGING) && !defined(NO_RCHECK) && !defined(NO_FILL_CHECK) && !defined(MALLOC_FILL_CHECK)
# define MALLOC_FILL_CHECK
#endif
#if defined(RCHECK) && defined(IGNORE_SMALL_BAD_FREE)
# undef IGNORE_SMALL_BAD_FREE
#endif
/*
* malloc.c (Caltech) 2/21/82
* Chris Kingsley, kingsley@cit-20.
*
* This is a very fast storage allocator. It allocates blocks of a small
* number of different sizes, and keeps free lists of each size. Blocks that
* don't exactly fit are passed up to the next larger size. In this
* implementation, the available sizes are 2^n-4 (or 2^n-12) bytes long.
* If PACK_MALLOC is defined, small blocks are 2^n bytes long.
* This is designed for use in a program that uses vast quantities of memory,
* but bombs when it runs out.
*
* Modifications Copyright Ilya Zakharevich 1996-99.
*
* Still very quick, but much more thrifty. (Std config is 10% slower
* than it was, and takes 67% of old heap size for typical usage.)
*
* Allocations of small blocks are now table-driven to many different
* buckets. Sizes of really big buckets are increased to accommodate
* common size=power-of-2 blocks. Running-out-of-memory is made into
* an exception. Deeply configurable and thread-safe.
*
*/
#include "EXTERN.h"
#define PERL_IN_MALLOC_C
#include "perl.h"
#if defined(MULTIPLICITY)
# define croak2 Perl_croak_nocontext
#else
# define croak2 croak
#endif
#ifdef USE_ITHREADS
# define PERL_MAYBE_ALIVE PL_thr_key
#else
# define PERL_MAYBE_ALIVE 1
#endif
#ifndef MYMALLOC
# error "MYMALLOC is not defined"
#endif
#ifndef MUTEX_LOCK
# define MUTEX_LOCK(l)
#endif
#ifndef MUTEX_UNLOCK
# define MUTEX_UNLOCK(l)
#endif
#ifndef MALLOC_LOCK
# define MALLOC_LOCK MUTEX_LOCK(&PL_malloc_mutex)
#endif
#ifndef MALLOC_UNLOCK
# define MALLOC_UNLOCK MUTEX_UNLOCK(&PL_malloc_mutex)
#endif
# ifndef fatalcroak /* make depend */
# define fatalcroak(mess) (write(2, (mess), strlen(mess)), exit(2))
# endif
#ifdef DEBUGGING
# undef DEBUG_m
# define DEBUG_m(a) \
STMT_START { \
if (PERL_MAYBE_ALIVE && PERL_GET_THX) { \
dTHX; \
if (DEBUG_m_TEST) { \
PL_debug &= ~DEBUG_m_FLAG; \
a; \
PL_debug |= DEBUG_m_FLAG; \
} \
} \
} STMT_END
#endif
#ifdef MULTIPLICITY
# define PERL_IS_ALIVE aTHX
#else
# define PERL_IS_ALIVE TRUE
#endif
/*
* Layout of memory:
* ~~~~~~~~~~~~~~~~
* The memory is broken into "blocks" which occupy multiples of 2K (and
* generally speaking, have size "close" to a power of 2). The addresses
* of such *unused* blocks are kept in nextf[i] with big enough i. (nextf
* is an array of linked lists.) (Addresses of used blocks are not known.)
*
* Moreover, since the algorithm may try to "bite" smaller blocks out
* of unused bigger ones, there are also regions of "irregular" size,
* managed separately, by a linked list chunk_chain.
*
* The third type of storage is the sbrk()ed-but-not-yet-used space, its
* end and size are kept in last_sbrk_top and sbrked_remains.
*
* Growing blocks "in place":
* ~~~~~~~~~~~~~~~~~~~~~~~~~
* The address of the block with the greatest address is kept in last_op
* (if not known, last_op is 0). If it is known that the memory above
* last_op is not continuous, or contains a chunk from chunk_chain,
* last_op is set to 0.
*
* The chunk with address last_op may be grown by expanding into
* sbrk()ed-but-not-yet-used space, or trying to sbrk() more continuous
* memory.
*
* Management of last_op:
* ~~~~~~~~~~~~~~~~~~~~~
*
* free() never changes the boundaries of blocks, so is not relevant.
*
* The only way realloc() may change the boundaries of blocks is if it
* grows a block "in place". However, in the case of success such a
* chunk is automatically last_op, and it remains last_op. In the case
* of failure getpages_adjacent() clears last_op.
*
* malloc() may change blocks by calling morecore() only.
*
* morecore() may create new blocks by:
* a) biting pieces from chunk_chain (cannot create one above last_op);
* b) biting a piece from an unused block (if block was last_op, this
* may create a chunk from chain above last_op, thus last_op is
* invalidated in such a case).
* c) biting of sbrk()ed-but-not-yet-used space. This creates
* a block which is last_op.
* d) Allocating new pages by calling getpages();
*
* getpages() creates a new block. It marks last_op at the bottom of
* the chunk of memory it returns.
*
* Active pages footprint:
* ~~~~~~~~~~~~~~~~~~~~~~
* Note that we do not need to traverse the lists in nextf[i], just take
* the first element of this list. However, we *need* to traverse the
* list in chunk_chain, but most the time it should be a very short one,
* so we do not step on a lot of pages we are not going to use.
*
* Flaws:
* ~~~~~
* get_from_bigger_buckets(): forget to increment price => Quite
* aggressive.
*/
/* I don't much care whether these are defined in sys/types.h--LAW */
#define u_char unsigned char
#define u_int unsigned int
/*
* I removed the definition of u_bigint which appeared to be u_bigint = UV
* u_bigint was only used in TWOK_MASKED and TWOK_SHIFT
* where I have used PTR2UV. RMB
*/
#define u_short unsigned short
#if defined(RCHECK) && defined(PACK_MALLOC)
# undef PACK_MALLOC
#endif
/*
* The description below is applicable if PACK_MALLOC is not defined.
*
* The overhead on a block is at least 4 bytes. When free, this space
* contains a pointer to the next free block, and the bottom two bits must
* be zero. When in use, the first byte is set to MAGIC, and the second
* byte is the size index. The remaining bytes are for alignment.
* If range checking is enabled and the size of the block fits
* in two bytes, then the top two bytes hold the size of the requested block
* plus the range checking words, and the header word MINUS ONE.
*/
union overhead {
union overhead *ov_next; /* when free */
#if MEM_ALIGNBYTES > 4
double strut; /* alignment problems */
# if MEM_ALIGNBYTES > 8
char sstrut[MEM_ALIGNBYTES]; /* for the sizing */
# endif
#endif
struct {
/*
* Keep the ovu_index and ovu_magic in this order, having a char
* field first gives alignment indigestion in some systems, such as
* MachTen.
*/
u_char ovu_index; /* bucket # */
u_char ovu_magic; /* magic number */
#ifdef RCHECK
/* Subtract one to fit into u_short for an extra bucket */
u_short ovu_size; /* block size (requested + overhead - 1) */
u_int ovu_rmagic; /* range magic number */
#endif
} ovu;
#define ov_magic ovu.ovu_magic
#define ov_index ovu.ovu_index
#define ov_size ovu.ovu_size
#define ov_rmagic ovu.ovu_rmagic
};
#define MAGIC 0xff /* magic # on accounting info */
#define RMAGIC 0x55555555 /* magic # on range info */
#define RMAGIC_C 0x55 /* magic # on range info */
#ifdef RCHECK
# define RMAGIC_SZ sizeof (u_int) /* Overhead at end of bucket */
# ifdef TWO_POT_OPTIMIZE
# define MAX_SHORT_BUCKET (12 * BUCKETS_PER_POW2) /* size-1 fits in short */
# else
# define MAX_SHORT_BUCKET (13 * BUCKETS_PER_POW2)
# endif
#else
# define RMAGIC_SZ 0
#endif
#if !defined(PACK_MALLOC) && defined(BUCKETS_ROOT2)
# undef BUCKETS_ROOT2
#endif
#ifdef BUCKETS_ROOT2
# define BUCKET_TABLE_SHIFT 2
# define BUCKET_POW2_SHIFT 1
# define BUCKETS_PER_POW2 2
#else
# define BUCKET_TABLE_SHIFT MIN_BUC_POW2
# define BUCKET_POW2_SHIFT 0
# define BUCKETS_PER_POW2 1
#endif
#if !defined(MEM_ALIGNBYTES) || ((MEM_ALIGNBYTES > 4) && !defined(STRICT_ALIGNMENT))
/* Figure out the alignment of void*. */
struct aligner {
char c;
void *p;
};
# define ALIGN_SMALL ((IV)((caddr_t)&(((struct aligner*)0)->p)))
#else
# define ALIGN_SMALL MEM_ALIGNBYTES
#endif
#define IF_ALIGN_8(yes,no) ((ALIGN_SMALL>4) ? (yes) : (no))
#ifdef BUCKETS_ROOT2
# define MAX_BUCKET_BY_TABLE 13
static const u_short buck_size[MAX_BUCKET_BY_TABLE + 1] =
{
0, 0, 0, 0, 4, 4, 8, 12, 16, 24, 32, 48, 64, 80,
};
# define BUCKET_SIZE_NO_SURPLUS(i) ((i) % 2 ? buck_size[i] : (1 << ((i) >> BUCKET_POW2_SHIFT)))
# define BUCKET_SIZE_REAL(i) ((i) <= MAX_BUCKET_BY_TABLE \
? ((size_t)buck_size[i]) \
: ((((size_t)1) << ((i) >> BUCKET_POW2_SHIFT)) \
- MEM_OVERHEAD(i) \
+ POW2_OPTIMIZE_SURPLUS(i)))
#else
# define BUCKET_SIZE_NO_SURPLUS(i) (((size_t)1) << ((i) >> BUCKET_POW2_SHIFT))
# define BUCKET_SIZE(i) (BUCKET_SIZE_NO_SURPLUS(i) + POW2_OPTIMIZE_SURPLUS(i))
# define BUCKET_SIZE_REAL(i) (BUCKET_SIZE(i) - MEM_OVERHEAD(i))
#endif
#ifdef PACK_MALLOC
/* In this case there are several possible layout of arenas depending
* on the size. Arenas are of sizes multiple to 2K, 2K-aligned, and
* have a size close to a power of 2.
*
* Arenas of the size >= 4K keep one chunk only. Arenas of size 2K
* may keep one chunk or multiple chunks. Here are the possible
* layouts of arenas:
*
* # One chunk only, chunksize 2^k + SOMETHING - ALIGN, k >= 11
*
* INDEX MAGIC1 UNUSED CHUNK1
*
* # Multichunk with sanity checking and chunksize 2^k-ALIGN, k>7
*
* INDEX MAGIC1 MAGIC2 MAGIC3 UNUSED CHUNK1 CHUNK2 CHUNK3 ...
*
* # Multichunk with sanity checking and size 2^k-ALIGN, k=7
*
* INDEX MAGIC1 MAGIC2 MAGIC3 UNUSED CHUNK1 UNUSED CHUNK2 CHUNK3 ...
*
* # Multichunk with sanity checking and size up to 80
*
* INDEX UNUSED MAGIC1 UNUSED MAGIC2 UNUSED ... CHUNK1 CHUNK2 CHUNK3 ...
*
* # No sanity check (usually up to 48=byte-long buckets)
* INDEX UNUSED CHUNK1 CHUNK2 ...
*
* Above INDEX and MAGIC are one-byte-long. Sizes of UNUSED are
* appropriate to keep algorithms simple and memory aligned. INDEX
* encodes the size of the chunk, while MAGICn encodes state (used,
* free or non-managed-by-us-so-it-indicates-a-bug) of CHUNKn. MAGIC
* is used for sanity checking purposes only. SOMETHING is 0 or 4K
* (to make size of big CHUNK accommodate allocations for powers of two
* better).
*
* [There is no need to alignment between chunks, since C rules ensure
* that structs which need 2^k alignment have sizeof which is
* divisible by 2^k. Thus as far as the last chunk is aligned at the
* end of the arena, and 2K-alignment does not contradict things,
* everything is going to be OK for sizes of chunks 2^n and 2^n +
* 2^k. Say, 80-bit buckets will be 16-bit aligned, and as far as we
* put allocations for requests in 65..80 range, all is fine.
*
* Note, however, that standard malloc() puts more strict
* requirements than the above C rules. Moreover, our algorithms of
* realloc() may break this idyll, but we suppose that realloc() does
* need not change alignment.]
*
* Is very important to make calculation of the offset of MAGICm as
* quick as possible, since it is done on each malloc()/free(). In
* fact it is so quick that it has quite little effect on the speed of
* doing malloc()/free(). [By default] We forego such calculations
* for small chunks, but only to save extra 3% of memory, not because
* of speed considerations.
*
* Here is the algorithm [which is the same for all the allocations
* schemes above], see OV_MAGIC(block,bucket). Let OFFSETm be the
* offset of the CHUNKm from the start of ARENA. Then offset of
* MAGICm is (OFFSET1 >> SHIFT) + ADDOFFSET. Here SHIFT and ADDOFFSET
* are numbers which depend on the size of the chunks only.
*
* Let as check some sanity conditions. Numbers OFFSETm>>SHIFT are
* different for all the chunks in the arena if 2^SHIFT is not greater
* than size of the chunks in the arena. MAGIC1 will not overwrite
* INDEX provided ADDOFFSET is >0 if OFFSET1 < 2^SHIFT. MAGIClast
* will not overwrite CHUNK1 if OFFSET1 > (OFFSETlast >> SHIFT) +
* ADDOFFSET.
*
* Make SHIFT the maximal possible (there is no point in making it
* smaller). Since OFFSETlast is 2K - CHUNKSIZE, above restrictions
* give restrictions on OFFSET1 and on ADDOFFSET.
*
* In particular, for chunks of size 2^k with k>=6 we can put
* ADDOFFSET to be from 0 to 2^k - 2^(11-k), and have
* OFFSET1==chunksize. For chunks of size 80 OFFSET1 of 2K%80=48 is
* large enough to have ADDOFFSET between 1 and 16 (similarly for 96,
* when ADDOFFSET should be 1). In particular, keeping MAGICs for
* these sizes gives no additional size penalty.
*
* However, for chunks of size 2^k with k<=5 this gives OFFSET1 >=
* ADDOFSET + 2^(11-k). Keeping ADDOFFSET 0 allows for 2^(11-k)-2^(11-2k)
* chunks per arena. This is smaller than 2^(11-k) - 1 which are
* needed if no MAGIC is kept. [In fact, having a negative ADDOFFSET
* would allow for slightly more buckets per arena for k=2,3.]
*
* Similarly, for chunks of size 3/2*2^k with k<=5 MAGICs would span
* the area up to 2^(11-k)+ADDOFFSET. For k=4 this give optimal
* ADDOFFSET as -7..0. For k=3 ADDOFFSET can go up to 4 (with tiny
* savings for negative ADDOFFSET). For k=5 ADDOFFSET can go -1..16
* (with no savings for negative values).
*
* In particular, keeping ADDOFFSET 0 for sizes of chunks up to 2^6
* leads to tiny pessimizations in case of sizes 4, 8, 12, 24, and
* leads to no contradictions except for size=80 (or 96.)
*
* However, it also makes sense to keep no magic for sizes 48 or less.
* This is what we do. In this case one needs ADDOFFSET>=1 also for
* chunksizes 12, 24, and 48, unless one gets one less chunk per
* arena.
*
* The algo of OV_MAGIC(block,bucket) keeps ADDOFFSET 0 until
* chunksize of 64, then makes it 1.
*
* This allows for an additional optimization: the above scheme leads
* to giant overheads for sizes 128 or more (one whole chunk needs to
* be sacrificed to keep INDEX). Instead we use chunks not of size
* 2^k, but of size 2^k-ALIGN. If we pack these chunks at the end of
* the arena, then the beginnings are still in different 2^k-long
* sections of the arena if k>=7 for ALIGN==4, and k>=8 if ALIGN=8.
* Thus for k>7 the above algo of calculating the offset of the magic
* will still give different answers for different chunks. And to
* avoid the overrun of MAGIC1 into INDEX, one needs ADDOFFSET of >=1.
* In the case k=7 we just move the first chunk an extra ALIGN
* backward inside the ARENA (this is done once per arena lifetime,
* thus is not a big overhead). */
# define MAX_PACKED_POW2 6
# define MAX_PACKED (MAX_PACKED_POW2 * BUCKETS_PER_POW2 + BUCKET_POW2_SHIFT)
# define MAX_POW2_ALGO ((1<<(MAX_PACKED_POW2 + 1)) - M_OVERHEAD)
# define TWOK_MASK nBIT_MASK(LOG_OF_MIN_ARENA)
# define TWOK_MASKED(x) (PTR2UV(x) & ~TWOK_MASK)
# define TWOK_SHIFT(x) (PTR2UV(x) & TWOK_MASK)
# define OV_INDEXp(block) (INT2PTR(u_char*,TWOK_MASKED(block)))
# define OV_INDEX(block) (*OV_INDEXp(block))
# define OV_MAGIC(block,bucket) (*(OV_INDEXp(block) + \
(TWOK_SHIFT(block)>> \
(bucket>>BUCKET_POW2_SHIFT)) + \
(bucket >= MIN_NEEDS_SHIFT ? 1 : 0)))
/* A bucket can have a shift smaller than it size, we need to
shift its magic number so it will not overwrite index: */
# ifdef BUCKETS_ROOT2
# define MIN_NEEDS_SHIFT (7*BUCKETS_PER_POW2 - 1) /* Shift 80 greater than chunk 64. */
# else
# define MIN_NEEDS_SHIFT (7*BUCKETS_PER_POW2) /* Shift 128 greater than chunk 32. */
# endif
# define CHUNK_SHIFT 0
/* Number of active buckets of given ordinal. */
#ifdef IGNORE_SMALL_BAD_FREE
#define FIRST_BUCKET_WITH_CHECK (6 * BUCKETS_PER_POW2) /* 64 */
# define N_BLKS(bucket) ( (bucket) < FIRST_BUCKET_WITH_CHECK \
? nBIT_MASK(LOG_OF_MIN_ARENA)/BUCKET_SIZE_NO_SURPLUS(bucket) \
: n_blks[bucket] )
#else
# define N_BLKS(bucket) n_blks[bucket]
#endif
static const u_short n_blks[LOG_OF_MIN_ARENA * BUCKETS_PER_POW2] =
{
# if BUCKETS_PER_POW2==1
0, 0,
(MIN_BUC_POW2==2 ? 384 : 0),
224, 120, 62, 31, 16, 8, 4, 2
# else
0, 0, 0, 0,
(MIN_BUC_POW2==2 ? 384 : 0), (MIN_BUC_POW2==2 ? 384 : 0), /* 4, 4 */
224, 149, 120, 80, 62, 41, 31, 25, 16, 16, 8, 8, 4, 4, 2, 2
# endif
};
/* Shift of the first bucket with the given ordinal inside 2K chunk. */
#ifdef IGNORE_SMALL_BAD_FREE
# define BLK_SHIFT(bucket) ( (bucket) < FIRST_BUCKET_WITH_CHECK \
? ((1<<LOG_OF_MIN_ARENA) \
- BUCKET_SIZE_NO_SURPLUS(bucket) * N_BLKS(bucket)) \
: blk_shift[bucket])
#else
# define BLK_SHIFT(bucket) blk_shift[bucket]
#endif
static const u_short blk_shift[LOG_OF_MIN_ARENA * BUCKETS_PER_POW2] =
{
# if BUCKETS_PER_POW2==1
0, 0,
(MIN_BUC_POW2==2 ? 512 : 0),
256, 128, 64, 64, /* 8 to 64 */
16*sizeof(union overhead),
8*sizeof(union overhead),
4*sizeof(union overhead),
2*sizeof(union overhead),
# else
0, 0, 0, 0,
(MIN_BUC_POW2==2 ? 512 : 0), (MIN_BUC_POW2==2 ? 512 : 0),
256, 260, 128, 128, 64, 80, 64, 48, /* 8 to 96 */
16*sizeof(union overhead), 16*sizeof(union overhead),
8*sizeof(union overhead), 8*sizeof(union overhead),
4*sizeof(union overhead), 4*sizeof(union overhead),
2*sizeof(union overhead), 2*sizeof(union overhead),
# endif
};
# define NEEDED_ALIGNMENT 0x800 /* 2k boundaries */
# define WANTED_ALIGNMENT 0x800 /* 2k boundaries */
#else /* !PACK_MALLOC */
# define OV_MAGIC(block,bucket) (block)->ov_magic
# define OV_INDEX(block) (block)->ov_index
# define CHUNK_SHIFT 1
# define MAX_PACKED -1
# define NEEDED_ALIGNMENT MEM_ALIGNBYTES
# define WANTED_ALIGNMENT 0x400 /* 1k boundaries */
#endif /* !PACK_MALLOC */
#define M_OVERHEAD (sizeof(union overhead) + RMAGIC_SZ) /* overhead at start+end */
#ifdef PACK_MALLOC
# define MEM_OVERHEAD(bucket) \
(bucket <= MAX_PACKED ? ((size_t)0) : M_OVERHEAD)
# ifdef SMALL_BUCKET_VIA_TABLE
# define START_SHIFTS_BUCKET ((MAX_PACKED_POW2 + 1) * BUCKETS_PER_POW2)
# define START_SHIFT MAX_PACKED_POW2
# ifdef BUCKETS_ROOT2 /* Chunks of size 3*2^n. */
# define SIZE_TABLE_MAX 80
# else
# define SIZE_TABLE_MAX 64
# endif
static const char bucket_of[] =
{
# ifdef BUCKETS_ROOT2 /* Chunks of size 3*2^n. */
/* 0 to 15 in 4-byte increments. */
(sizeof(void*) > 4 ? 6 : 5), /* 4/8, 5-th bucket for better reports */
6, /* 8 */
IF_ALIGN_8(8,7), 8, /* 16/12, 16 */
9, 9, 10, 10, /* 24, 32 */
11, 11, 11, 11, /* 48 */
12, 12, 12, 12, /* 64 */
13, 13, 13, 13, /* 80 */
13, 13, 13, 13 /* 80 */
# else /* !BUCKETS_ROOT2 */
/* 0 to 15 in 4-byte increments. */
(sizeof(void*) > 4 ? 3 : 2),
3,
4, 4,
5, 5, 5, 5,
6, 6, 6, 6,
6, 6, 6, 6
# endif /* !BUCKETS_ROOT2 */
};
# else /* !SMALL_BUCKET_VIA_TABLE */
# define START_SHIFTS_BUCKET MIN_BUCKET
# define START_SHIFT (MIN_BUC_POW2 - 1)
# endif /* !SMALL_BUCKET_VIA_TABLE */
#else /* !PACK_MALLOC */
# define MEM_OVERHEAD(bucket) M_OVERHEAD
# ifdef SMALL_BUCKET_VIA_TABLE
# undef SMALL_BUCKET_VIA_TABLE
# endif
# define START_SHIFTS_BUCKET MIN_BUCKET
# define START_SHIFT (MIN_BUC_POW2 - 1)
#endif /* !PACK_MALLOC */
/*
* Big allocations are often of the size 2^n bytes. To make them a
* little bit better, make blocks of size 2^n+pagesize for big n.
*/
#ifdef TWO_POT_OPTIMIZE
# ifndef PERL_PAGESIZE
# define PERL_PAGESIZE 4096
# endif
# ifndef FIRST_BIG_POW2
# define FIRST_BIG_POW2 15 /* 32K, 16K is used too often. */
# endif
# define FIRST_BIG_BLOCK (1<<FIRST_BIG_POW2)
/* If this value or more, check against bigger blocks. */
# define FIRST_BIG_BOUND (FIRST_BIG_BLOCK - M_OVERHEAD)
/* If less than this value, goes into 2^n-overhead-block. */
# define LAST_SMALL_BOUND ((FIRST_BIG_BLOCK>>1) - M_OVERHEAD)
# define POW2_OPTIMIZE_ADJUST(nbytes) \
((nbytes >= FIRST_BIG_BOUND) ? nbytes -= PERL_PAGESIZE : 0)
# define POW2_OPTIMIZE_SURPLUS(bucket) \
((size_t)((bucket >= FIRST_BIG_POW2 * BUCKETS_PER_POW2) ? PERL_PAGESIZE : 0))
#else /* !TWO_POT_OPTIMIZE */
# define POW2_OPTIMIZE_ADJUST(nbytes)
# define POW2_OPTIMIZE_SURPLUS(bucket) ((size_t)0)
#endif /* !TWO_POT_OPTIMIZE */
#define BARK_64K_LIMIT(what,nbytes,size)
#ifndef MIN_SBRK
# define MIN_SBRK 2048
#endif
#ifndef FIRST_SBRK
# define FIRST_SBRK (48*1024)
#endif
/* Minimal sbrk in percents of what is already alloced. */
#ifndef MIN_SBRK_FRAC
# define MIN_SBRK_FRAC 3
#endif
#ifndef SBRK_ALLOW_FAILURES
# define SBRK_ALLOW_FAILURES 3
#endif
#ifndef SBRK_FAILURE_PRICE
# define SBRK_FAILURE_PRICE 50
#endif
static void morecore (int bucket);
# if defined(DEBUGGING)
static void botch (const char *diag, const char *s, const char *file, int line);
# endif
static void add_to_chain (void *p, MEM_SIZE size, MEM_SIZE chip);
static void* get_from_chain (MEM_SIZE size);
static void* get_from_bigger_buckets(int bucket, MEM_SIZE size);
static union overhead *getpages (MEM_SIZE needed, int *nblksp, int bucket);
static int getpages_adjacent(MEM_SIZE require);
#ifdef I_MACH_CTHREADS
# undef MUTEX_LOCK
# define MUTEX_LOCK(m) STMT_START { if (*m) mutex_lock(*m); } STMT_END
# undef MUTEX_UNLOCK
# define MUTEX_UNLOCK(m) STMT_START { if (*m) mutex_unlock(*m); } STMT_END
#endif
#ifndef PTRSIZE
# define PTRSIZE sizeof(void*)
#endif
#ifndef BITS_IN_PTR
# define BITS_IN_PTR (8*PTRSIZE)
#endif
/*
* nextf[i] is the pointer to the next free block of size 2^i. The
* smallest allocatable block is 8 bytes. The overhead information
* precedes the data area returned to the user.
*/
#define NBUCKETS (BITS_IN_PTR*BUCKETS_PER_POW2 + 1)
static union overhead *nextf[NBUCKETS];
#if defined(PURIFY) && !defined(USE_PERL_SBRK)
# define USE_PERL_SBRK
#endif
#ifdef USE_PERL_SBRK
# define sbrk(a) Perl_sbrk(a)
Malloc_t Perl_sbrk (int size);
#elif !defined(HAS_SBRK_PROTO) /* <unistd.h> usually takes care of this */
extern Malloc_t sbrk(int);
#endif
#ifndef MIN_SBRK_FRAC1000 /* Backward compatibility */
# define MIN_SBRK_FRAC1000 (MIN_SBRK_FRAC * 10)
#endif
#include "malloc_ctl.h"
#ifndef NO_MALLOC_DYNAMIC_CFG
# define PERL_MALLOC_OPT_CHARS "FMfAPGdac"
# ifndef FILL_DEAD_DEFAULT
# define FILL_DEAD_DEFAULT 1
# endif
# ifndef FILL_ALIVE_DEFAULT
# define FILL_ALIVE_DEFAULT 1
# endif
# ifndef FILL_CHECK_DEFAULT
# define FILL_CHECK_DEFAULT 1
# endif
static IV MallocCfg[MallocCfg_last] = {
FIRST_SBRK,
MIN_SBRK,
MIN_SBRK_FRAC,
SBRK_ALLOW_FAILURES,
SBRK_FAILURE_PRICE,
SBRK_ALLOW_FAILURES * SBRK_FAILURE_PRICE, /* sbrk_goodness */
FILL_DEAD_DEFAULT, /* FILL_DEAD */
FILL_ALIVE_DEFAULT, /* FILL_ALIVE */
FILL_CHECK_DEFAULT, /* FILL_CHECK */
0, /* MallocCfg_skip_cfg_env */
0, /* MallocCfg_cfg_env_read */
0, /* MallocCfg_emergency_buffer_size */
0, /* MallocCfg_emergency_buffer_prepared_size */
0 /* MallocCfg_emergency_buffer_last_req */
};
IV *MallocCfg_ptr = MallocCfg;
static char* MallocCfgP[MallocCfg_last] = {
0, /* MallocCfgP_emergency_buffer */
0, /* MallocCfgP_emergency_buffer_prepared */
};
char **MallocCfgP_ptr = MallocCfgP;
# undef MIN_SBRK
# undef FIRST_SBRK
# undef MIN_SBRK_FRAC1000
# undef SBRK_ALLOW_FAILURES
# undef SBRK_FAILURE_PRICE
# define MIN_SBRK MallocCfg[MallocCfg_MIN_SBRK]
# define FIRST_SBRK MallocCfg[MallocCfg_FIRST_SBRK]
# define MIN_SBRK_FRAC1000 MallocCfg[MallocCfg_MIN_SBRK_FRAC1000]
# define SBRK_ALLOW_FAILURES MallocCfg[MallocCfg_SBRK_ALLOW_FAILURES]
# define SBRK_FAILURE_PRICE MallocCfg[MallocCfg_SBRK_FAILURE_PRICE]
# define sbrk_goodness MallocCfg[MallocCfg_sbrk_goodness]
# define emergency_buffer_size MallocCfg[MallocCfg_emergency_buffer_size]
# define emergency_buffer_last_req MallocCfg[MallocCfg_emergency_buffer_last_req]
# define FILL_DEAD MallocCfg[MallocCfg_filldead]
# define FILL_ALIVE MallocCfg[MallocCfg_fillalive]
# define FILL_CHECK_CFG MallocCfg[MallocCfg_fillcheck]
# define FILL_CHECK (FILL_DEAD && FILL_CHECK_CFG)
# define emergency_buffer MallocCfgP[MallocCfgP_emergency_buffer]
# define emergency_buffer_prepared MallocCfgP[MallocCfgP_emergency_buffer_prepared]
#else /* defined(NO_MALLOC_DYNAMIC_CFG) */
# define FILL_DEAD 1
# define FILL_ALIVE 1
# define FILL_CHECK 1
static int sbrk_goodness = SBRK_ALLOW_FAILURES * SBRK_FAILURE_PRICE;
# define NO_PERL_MALLOC_ENV
#endif
#ifdef DEBUGGING_MSTATS
/*
* nmalloc[i] is the difference between the number of mallocs and frees
* for a given block size.
*/
static u_int nmalloc[NBUCKETS];
static u_int sbrk_slack;
static u_int start_slack;
#else /* !( defined DEBUGGING_MSTATS ) */
# define sbrk_slack 0
#endif
static u_int goodsbrk;
#ifdef PERL_EMERGENCY_SBRK
# ifndef BIG_SIZE
# define BIG_SIZE (1<<16) /* 64K */
# endif
# ifdef NO_MALLOC_DYNAMIC_CFG
static MEM_SIZE emergency_buffer_size;
/* 0 if the last request for more memory succeeded.
Otherwise the size of the failing request. */
static MEM_SIZE emergency_buffer_last_req;
static char *emergency_buffer;
static char *emergency_buffer_prepared;
# endif
# ifndef emergency_sbrk_croak
# define emergency_sbrk_croak croak2
# endif
static char *
perl_get_emergency_buffer(IV *size)
{
dTHX;
/* First offense, give a possibility to recover by dieing. */
/* No malloc involved here: */
SV *sv;
char *pv;
GV **gvp = (GV**)hv_fetchs(PL_defstash, "^M", FALSE);
if (!gvp) gvp = (GV**)hv_fetchs(PL_defstash, "\015", FALSE);
if (!gvp || !(sv = GvSV(*gvp)) || !SvPOK(sv)
|| (SvLEN(sv) < (1<<LOG_OF_MIN_ARENA) - M_OVERHEAD))
return NULL; /* Now die die die... */
/* Got it, now detach SvPV: */
pv = SvPV_nolen(sv);
/* Check alignment: */
if ((PTR2UV(pv) - sizeof(union overhead)) & (NEEDED_ALIGNMENT - 1)) {
PerlIO_puts(PerlIO_stderr(),"Bad alignment of $^M!\n");
return NULL; /* die die die */
}
SvPOK_off(sv);
SvPV_set(sv, NULL);
SvCUR_set(sv, 0);
SvLEN_set(sv, 0);
*size = malloced_size(pv) + M_OVERHEAD;
return pv - sizeof(union overhead);
}
# define PERL_GET_EMERGENCY_BUFFER(p) perl_get_emergency_buffer(p)
# ifndef NO_MALLOC_DYNAMIC_CFG
static char *
get_emergency_buffer(IV *size)
{
char *pv = emergency_buffer_prepared;
*size = MallocCfg[MallocCfg_emergency_buffer_prepared_size];
emergency_buffer_prepared = 0;
MallocCfg[MallocCfg_emergency_buffer_prepared_size] = 0;
return pv;
}
# define GET_EMERGENCY_BUFFER(p) get_emergency_buffer(p)
# else /* NO_MALLOC_DYNAMIC_CFG */
# define GET_EMERGENCY_BUFFER(p) NULL
# endif
static Malloc_t
emergency_sbrk(MEM_SIZE size)
{
MEM_SIZE rsize = (((size - 1)>>LOG_OF_MIN_ARENA) + 1)<<LOG_OF_MIN_ARENA;
if (size >= BIG_SIZE
&& (!emergency_buffer_last_req ||
(size < (MEM_SIZE)emergency_buffer_last_req))) {
/* Give the possibility to recover, but avoid an infinite cycle. */
MALLOC_UNLOCK;
emergency_buffer_last_req = size;
emergency_sbrk_croak("Out of memory during \"large\" request for %" UVuf
" bytes, total sbrk() is %" UVuf " bytes",
(UV)size, (UV)(goodsbrk + sbrk_slack));
}