Skip to content
Projects
Groups
Snippets
Help
This project
Loading...
Sign in / Register
Toggle navigation
W
wine-cw
Project
Project
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Registry
Registry
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
wine
wine-cw
Commits
0044e527
Commit
0044e527
authored
May 27, 2022
by
Rémi Bernon
Committed by
Alexandre Julliard
May 27, 2022
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
ntdll: Use struct entry directly for the free lists.
Signed-off-by:
Rémi Bernon
<
rbernon@codeweavers.com
>
parent
2cc6f00a
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
22 additions
and
25 deletions
+22
-25
heap.c
dlls/ntdll/heap.c
+22
-25
No files found.
dlls/ntdll/heap.c
View file @
0044e527
...
...
@@ -74,6 +74,8 @@ struct rtl_heap_entry
/* header for heap blocks */
#define ALIGNMENT (2 * sizeof(void *))
typedef
struct
block
{
DWORD
size
;
/* Block size; must be the first field */
...
...
@@ -92,11 +94,13 @@ C_ASSERT( sizeof(struct block) == 8 );
/* entry to link free blocks in free lists */
typedef
struct
entry
struct
DECLSPEC_ALIGN
(
ALIGNMENT
)
entry
{
struct
block
block
;
struct
list
entry
;
}
ARENA_FREE
;
};
C_ASSERT
(
sizeof
(
struct
entry
)
==
2
*
ALIGNMENT
);
typedef
struct
{
...
...
@@ -122,7 +126,6 @@ typedef struct
#define ARENA_FREE_FILLER 0xfeeefeee
/* everything is aligned on 8 byte boundaries (16 for Win64) */
#define ALIGNMENT (2*sizeof(void*))
#define LARGE_ALIGNMENT 16
/* large blocks have stricter alignment */
#define ARENA_OFFSET (ALIGNMENT - sizeof(ARENA_INUSE))
#define COMMIT_MASK 0xffff
/* bitmask for commit/decommit granularity */
...
...
@@ -137,7 +140,7 @@ C_ASSERT( sizeof(ARENA_LARGE) % LARGE_ALIGNMENT == 0 );
#define HEAP_MIN_DATA_SIZE ROUND_SIZE(2 * sizeof(struct list))
#define HEAP_MIN_BLOCK_SIZE (HEAP_MIN_DATA_SIZE + sizeof(ARENA_INUSE))
/* minimum size that must remain to shrink an allocated block */
#define HEAP_MIN_SHRINK_SIZE (HEAP_MIN_DATA_SIZE+sizeof(
ARENA_FREE
))
#define HEAP_MIN_SHRINK_SIZE (HEAP_MIN_DATA_SIZE+sizeof(
struct entry
))
/* minimum size to start allocating large blocks */
#define HEAP_MIN_LARGE_BLOCK_SIZE (0x10000 * ALIGNMENT - 0x1000)
/* extra size to add at the end of block for tail checking */
...
...
@@ -156,12 +159,6 @@ static const SIZE_T free_list_sizes[] =
};
#define HEAP_NB_FREE_LISTS (ARRAY_SIZE(free_list_sizes) + HEAP_NB_SMALL_FREE_LISTS)
typedef
union
{
ARENA_FREE
arena
;
void
*
alignment
[
4
];
}
FREE_LIST_ENTRY
;
struct
tagHEAP
;
typedef
struct
DECLSPEC_ALIGN
(
ALIGNMENT
)
tagSUBHEAP
...
...
@@ -200,7 +197,7 @@ typedef struct tagHEAP
DWORD
pending_pos
;
/* Position in pending free requests ring */
ARENA_INUSE
**
pending_free
;
/* Ring buffer for pending free requests */
RTL_CRITICAL_SECTION
cs
;
FREE_LIST_ENTRY
free_lists
[
HEAP_NB_FREE_LISTS
];
struct
entry
free_lists
[
HEAP_NB_FREE_LISTS
];
SUBHEAP
subheap
;
}
HEAP
;
...
...
@@ -249,7 +246,7 @@ static inline void block_set_type( struct block *block, UINT type )
static
inline
UINT
block_get_overhead
(
const
struct
block
*
block
)
{
if
(
block_get_flags
(
block
)
&
BLOCK_FLAG_FREE
)
return
sizeof
(
struct
entry
);
if
(
block_get_flags
(
block
)
&
BLOCK_FLAG_FREE
)
return
sizeof
(
*
block
)
+
sizeof
(
struct
list
);
return
sizeof
(
*
block
)
+
block
->
unused_bytes
;
}
...
...
@@ -425,7 +422,7 @@ static void notify_free_all( SUBHEAP *subheap )
/* size is the size of the whole block including the arena header */
static
inline
struct
entry
*
find_free_list
(
HEAP
*
heap
,
SIZE_T
block_size
,
BOOL
last
)
{
FREE_LIST_ENTRY
*
list
,
*
end
=
heap
->
free_lists
+
ARRAY_SIZE
(
heap
->
free_lists
);
struct
entry
*
list
,
*
end
=
heap
->
free_lists
+
ARRAY_SIZE
(
heap
->
free_lists
);
unsigned
int
i
;
if
(
block_size
<=
HEAP_MAX_SMALL_FREE_LIST
)
...
...
@@ -435,7 +432,7 @@ static inline struct entry *find_free_list( HEAP *heap, SIZE_T block_size, BOOL
list
=
heap
->
free_lists
+
i
;
if
(
last
&&
++
list
==
end
)
list
=
heap
->
free_lists
;
return
&
list
->
arena
;
return
list
;
}
/* get the memory protection type to use for a given heap */
...
...
@@ -492,8 +489,8 @@ static void heap_dump( const HEAP *heap )
if
(
i
<
HEAP_NB_SMALL_FREE_LISTS
)
size
=
HEAP_MIN_BLOCK_SIZE
+
i
*
ALIGNMENT
;
else
size
=
free_list_sizes
[
i
-
HEAP_NB_SMALL_FREE_LISTS
];
TRACE
(
" %p: size %8Ix, prev %p, next %p
\n
"
,
heap
->
free_lists
+
i
,
size
,
LIST_ENTRY
(
heap
->
free_lists
[
i
].
arena
.
entry
.
prev
,
struct
entry
,
entry
),
LIST_ENTRY
(
heap
->
free_lists
[
i
].
arena
.
entry
.
next
,
struct
entry
,
entry
)
);
LIST_ENTRY
(
heap
->
free_lists
[
i
].
entry
.
prev
,
struct
entry
,
entry
),
LIST_ENTRY
(
heap
->
free_lists
[
i
].
entry
.
next
,
struct
entry
,
entry
)
);
}
TRACE
(
" subheaps: %p
\n
"
,
&
heap
->
subheap_list
);
...
...
@@ -629,7 +626,7 @@ static inline BOOL subheap_commit( SUBHEAP *subheap, const struct block *block,
SIZE_T
size
;
void
*
addr
;
commit_end
=
(
char
*
)
block
+
block_size
+
sizeof
(
ARENA_FREE
);
commit_end
=
(
char
*
)
block
+
block_size
+
sizeof
(
struct
entry
);
commit_end
=
ROUND_ADDR
((
char
*
)
commit_end
+
COMMIT_MASK
,
COMMIT_MASK
);
if
(
commit_end
>
end
)
commit_end
=
end
;
...
...
@@ -908,8 +905,8 @@ static BOOL validate_large_arena( const HEAP *heap, const ARENA_LARGE *arena )
static
SUBHEAP
*
HEAP_CreateSubHeap
(
HEAP
*
heap
,
LPVOID
address
,
DWORD
flags
,
SIZE_T
commitSize
,
SIZE_T
totalSize
)
{
struct
entry
*
pEntry
;
SUBHEAP
*
subheap
;
FREE_LIST_ENTRY
*
pEntry
;
unsigned
int
i
;
if
(
!
address
)
...
...
@@ -968,12 +965,12 @@ static SUBHEAP *HEAP_CreateSubHeap( HEAP *heap, LPVOID address, DWORD flags,
/* Build the free lists */
list_init
(
&
heap
->
free_lists
[
0
].
arena
.
entry
);
list_init
(
&
heap
->
free_lists
[
0
].
entry
);
for
(
i
=
0
,
pEntry
=
heap
->
free_lists
;
i
<
HEAP_NB_FREE_LISTS
;
i
++
,
pEntry
++
)
{
block_set_size
(
&
pEntry
->
arena
.
block
,
BLOCK_FLAG_FREE_LINK
,
0
);
block_set_type
(
&
pEntry
->
arena
.
block
,
ARENA_FREE_MAGIC
);
if
(
i
)
list_add_after
(
&
pEntry
[
-
1
].
arena
.
entry
,
&
pEntry
->
arena
.
entry
);
block_set_size
(
&
pEntry
->
block
,
BLOCK_FLAG_FREE_LINK
,
0
);
block_set_type
(
&
pEntry
->
block
,
ARENA_FREE_MAGIC
);
if
(
i
)
list_add_after
(
&
pEntry
[
-
1
].
entry
,
&
pEntry
->
entry
);
}
/* Initialize critical section */
...
...
@@ -1025,7 +1022,7 @@ static struct block *find_free_block( HEAP *heap, SIZE_T block_size, SUBHEAP **s
/* Find a suitable free list, and in it find a block large enough */
while
((
ptr
=
list_next
(
&
heap
->
free_lists
[
0
].
arena
.
entry
,
ptr
)))
while
((
ptr
=
list_next
(
&
heap
->
free_lists
[
0
].
entry
,
ptr
)))
{
entry
=
LIST_ENTRY
(
ptr
,
struct
entry
,
entry
);
block
=
(
struct
block
*
)
entry
;
...
...
@@ -1048,7 +1045,7 @@ static struct block *find_free_block( HEAP *heap, SIZE_T block_size, SUBHEAP **s
}
/* make sure we can fit the block and a free entry at the end */
total_size
=
sizeof
(
SUBHEAP
)
+
block_size
+
sizeof
(
ARENA_FREE
);
total_size
=
sizeof
(
SUBHEAP
)
+
block_size
+
sizeof
(
struct
entry
);
if
(
total_size
<
block_size
)
return
NULL
;
/* overflow */
if
((
*
subheap
=
HEAP_CreateSubHeap
(
heap
,
NULL
,
heap
->
flags
,
total_size
,
...
...
@@ -1078,7 +1075,7 @@ static BOOL is_valid_free_block( const HEAP *heap, const struct block *block )
unsigned
int
i
;
if
((
subheap
=
find_subheap
(
heap
,
block
,
FALSE
)))
return
TRUE
;
for
(
i
=
0
;
i
<
HEAP_NB_FREE_LISTS
;
i
++
)
if
(
block
==
&
heap
->
free_lists
[
i
].
arena
.
block
)
return
TRUE
;
for
(
i
=
0
;
i
<
HEAP_NB_FREE_LISTS
;
i
++
)
if
(
block
==
&
heap
->
free_lists
[
i
].
block
)
return
TRUE
;
return
FALSE
;
}
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment