@@ -511,7 +511,7 @@ INTERNAL_HIDDEN iso_alloc_zone_t *_iso_new_zone(size_t size, bool internal, int3
511511 * but when we call create_canary_chunks() that will happen anyway */
512512 p = mmap_rw_pages (total_size , false, name );
513513
514- #if NAMED_MAPPINGS && __ANDROID__
514+ #if __ANDROID__ && NAMED_MAPPINGS && MEMORY_TAGGING
515515 if (new_zone -> tagged == false) {
516516 name = MEM_TAG_NAME ;
517517 }
@@ -1205,15 +1205,13 @@ INTERNAL_HIDDEN iso_alloc_zone_t *iso_find_zone_bitmap_range(const void *restric
12051205 return zone ;
12061206 }
12071207
1208- iso_alloc_zone_t * tmp_zone = NULL ;
1209-
12101208 /* Now we check the MRU thread zone cache */
12111209 for (int64_t i = 0 ; i < zone_cache_count ; i ++ ) {
1212- tmp_zone = zone_cache [i ].zone ;
1213- bitmap_start = UNMASK_BITMAP_PTR (tmp_zone );
1210+ zone = zone_cache [i ].zone ;
1211+ bitmap_start = UNMASK_BITMAP_PTR (zone );
12141212
1215- if (bitmap_start <= p && (bitmap_start + tmp_zone -> bitmap_size ) > p ) {
1216- return tmp_zone ;
1213+ if (bitmap_start <= p && (bitmap_start + zone -> bitmap_size ) > p ) {
1214+ return zone ;
12171215 }
12181216 }
12191217
@@ -1240,15 +1238,13 @@ INTERNAL_HIDDEN iso_alloc_zone_t *iso_find_zone_range(const void *restrict p) {
12401238 return zone ;
12411239 }
12421240
1243- iso_alloc_zone_t * tmp_zone = NULL ;
1244-
12451241 /* Now we check the MRU thread zone cache */
12461242 for (int64_t i = 0 ; i < zone_cache_count ; i ++ ) {
1247- tmp_zone = zone_cache [i ].zone ;
1248- user_pages_start = UNMASK_USER_PTR (tmp_zone );
1243+ zone = zone_cache [i ].zone ;
1244+ user_pages_start = UNMASK_USER_PTR (zone );
12491245
12501246 if (user_pages_start <= p && (user_pages_start + ZONE_USER_SIZE ) > p ) {
1251- return tmp_zone ;
1247+ return zone ;
12521248 }
12531249 }
12541250
0 commit comments