Angel Pons | 118a9c7 | 2020-04-02 23:48:34 +0200 | [diff] [blame] | 1 | /* SPDX-License-Identifier: GPL-2.0-only */ |
Aaron Durbin | 0424c95 | 2015-03-28 23:56:22 -0500 | [diff] [blame] | 2 | |
| 3 | #include <boot_device.h> |
Julius Werner | cefe89e | 2019-11-06 19:29:44 -0800 | [diff] [blame] | 4 | #include <cbmem.h> |
Aaron Durbin | 0424c95 | 2015-03-28 23:56:22 -0500 | [diff] [blame] | 5 | #include <console/console.h> |
| 6 | #include <fmap.h> |
Julius Werner | fdabf3f | 2020-05-06 17:06:35 -0700 | [diff] [blame] | 7 | #include <metadata_hash.h> |
Aaron Durbin | 0424c95 | 2015-03-28 23:56:22 -0500 | [diff] [blame] | 8 | #include <stddef.h> |
| 9 | #include <string.h> |
Julius Werner | cefe89e | 2019-11-06 19:29:44 -0800 | [diff] [blame] | 10 | #include <symbols.h> |
Krystian Hebel | 93f6b8a | 2020-09-30 18:23:14 +0200 | [diff] [blame] | 11 | #include <endian.h> |
Aaron Durbin | 0424c95 | 2015-03-28 23:56:22 -0500 | [diff] [blame] | 12 | |
Aaron Durbin | bf1e481 | 2016-05-10 15:12:08 -0500 | [diff] [blame] | 13 | #include "fmap_config.h" |
| 14 | |
Aaron Durbin | 0424c95 | 2015-03-28 23:56:22 -0500 | [diff] [blame] | 15 | /* |
| 16 | * See http://code.google.com/p/flashmap/ for more information on FMAP. |
| 17 | */ |
| 18 | |
Arthur Heymans | dba22d2 | 2019-11-20 19:57:49 +0100 | [diff] [blame] | 19 | static int fmap_print_once; |
Julius Werner | c893197 | 2021-04-16 16:48:32 -0700 | [diff] [blame] | 20 | static struct region_device fmap_cache; |
Duncan Laurie | bc2c0a3 | 2016-02-09 09:17:56 -0800 | [diff] [blame] | 21 | |
Julius Werner | cefe89e | 2019-11-06 19:29:44 -0800 | [diff] [blame] | 22 | #define print_once(...) do { \ |
Arthur Heymans | dba22d2 | 2019-11-20 19:57:49 +0100 | [diff] [blame] | 23 | if (!fmap_print_once) \ |
Julius Werner | cefe89e | 2019-11-06 19:29:44 -0800 | [diff] [blame] | 24 | printk(__VA_ARGS__); \ |
| 25 | } while (0) |
| 26 | |
Furquan Shaikh | b33a2b0 | 2019-09-26 23:51:46 -0700 | [diff] [blame] | 27 | uint64_t get_fmap_flash_offset(void) |
| 28 | { |
| 29 | return FMAP_OFFSET; |
| 30 | } |
| 31 | |
Julius Werner | fdabf3f | 2020-05-06 17:06:35 -0700 | [diff] [blame] | 32 | static int verify_fmap(const struct fmap *fmap) |
Julius Werner | cefe89e | 2019-11-06 19:29:44 -0800 | [diff] [blame] | 33 | { |
Julius Werner | ca71588 | 2023-11-02 16:20:17 -0700 | [diff] [blame] | 34 | if (memcmp(fmap->signature, FMAP_SIGNATURE, sizeof(fmap->signature))) { |
| 35 | if (ENV_INITIAL_STAGE) |
| 36 | printk(BIOS_ERR, "Invalid FMAP at %#x\n", FMAP_OFFSET); |
Julius Werner | fdabf3f | 2020-05-06 17:06:35 -0700 | [diff] [blame] | 37 | return -1; |
Julius Werner | ca71588 | 2023-11-02 16:20:17 -0700 | [diff] [blame] | 38 | } |
Julius Werner | fdabf3f | 2020-05-06 17:06:35 -0700 | [diff] [blame] | 39 | |
| 40 | static bool done = false; |
| 41 | if (!CONFIG(CBFS_VERIFICATION) || !ENV_INITIAL_STAGE || done) |
| 42 | return 0; /* Only need to check hash in first stage. */ |
| 43 | |
Julius Werner | 682cb3b | 2023-11-02 15:44:12 -0700 | [diff] [blame] | 44 | /* On error we need to die right here, lest we risk a TOCTOU attack where the cache is |
| 45 | filled with a tampered FMAP but the later fallback path is fed a valid one. */ |
Julius Werner | fdabf3f | 2020-05-06 17:06:35 -0700 | [diff] [blame] | 46 | if (metadata_hash_verify_fmap(fmap, FMAP_SIZE) != VB2_SUCCESS) |
Julius Werner | 682cb3b | 2023-11-02 15:44:12 -0700 | [diff] [blame] | 47 | die("FMAP verification failure"); |
Julius Werner | fdabf3f | 2020-05-06 17:06:35 -0700 | [diff] [blame] | 48 | |
| 49 | done = true; |
| 50 | return 0; |
Julius Werner | cefe89e | 2019-11-06 19:29:44 -0800 | [diff] [blame] | 51 | } |
| 52 | |
| 53 | static void report(const struct fmap *fmap) |
| 54 | { |
| 55 | print_once(BIOS_DEBUG, "FMAP: Found \"%s\" version %d.%d at %#x.\n", |
| 56 | fmap->name, fmap->ver_major, fmap->ver_minor, FMAP_OFFSET); |
| 57 | print_once(BIOS_DEBUG, "FMAP: base = %#llx size = %#x #areas = %d\n", |
Krystian Hebel | 93f6b8a | 2020-09-30 18:23:14 +0200 | [diff] [blame] | 58 | (long long)le64toh(fmap->base), le32toh(fmap->size), |
| 59 | le16toh(fmap->nareas)); |
Arthur Heymans | dba22d2 | 2019-11-20 19:57:49 +0100 | [diff] [blame] | 60 | fmap_print_once = 1; |
Julius Werner | cefe89e | 2019-11-06 19:29:44 -0800 | [diff] [blame] | 61 | } |
| 62 | |
Julius Werner | c893197 | 2021-04-16 16:48:32 -0700 | [diff] [blame] | 63 | static void setup_preram_cache(struct region_device *cache_rdev) |
Julius Werner | cefe89e | 2019-11-06 19:29:44 -0800 | [diff] [blame] | 64 | { |
Julius Werner | 7fc9286 | 2019-11-18 13:01:06 -0800 | [diff] [blame] | 65 | if (CONFIG(NO_FMAP_CACHE)) |
| 66 | return; |
| 67 | |
Josie Nordrum | c3cc158 | 2020-09-09 12:57:13 -0600 | [diff] [blame] | 68 | /* No need to use FMAP cache in SMM */ |
| 69 | if (ENV_SMM) |
| 70 | return; |
| 71 | |
Julius Werner | cefe89e | 2019-11-06 19:29:44 -0800 | [diff] [blame] | 72 | if (!ENV_ROMSTAGE_OR_BEFORE) { |
| 73 | /* We get here if ramstage makes an FMAP access before calling |
| 74 | cbmem_initialize(). We should avoid letting it come to that, |
| 75 | so print a warning. */ |
| 76 | print_once(BIOS_WARNING, |
| 77 | "WARNING: Post-RAM FMAP access too early for cache!\n"); |
| 78 | return; |
| 79 | } |
| 80 | |
Julius Werner | cefe89e | 2019-11-06 19:29:44 -0800 | [diff] [blame] | 81 | struct fmap *fmap = (struct fmap *)_fmap_cache; |
Martin Roth | 1594e8f | 2020-07-15 13:57:54 -0600 | [diff] [blame] | 82 | if (!(ENV_INITIAL_STAGE)) { |
| 83 | /* NOTE: This assumes that the first stage will make |
Julius Werner | cefe89e | 2019-11-06 19:29:44 -0800 | [diff] [blame] | 84 | at least one FMAP access (usually from finding CBFS). */ |
Julius Werner | fdabf3f | 2020-05-06 17:06:35 -0700 | [diff] [blame] | 85 | if (!verify_fmap(fmap)) |
Julius Werner | cefe89e | 2019-11-06 19:29:44 -0800 | [diff] [blame] | 86 | goto register_cache; |
| 87 | |
Julius Werner | ca71588 | 2023-11-02 16:20:17 -0700 | [diff] [blame] | 88 | /* This shouldn't happen, so no point providing a fallback path here. */ |
| 89 | die("FMAP cache corrupted?!\n"); |
Julius Werner | cefe89e | 2019-11-06 19:29:44 -0800 | [diff] [blame] | 90 | } |
| 91 | |
| 92 | /* In case we fail below, make sure the cache is invalid. */ |
| 93 | memset(fmap->signature, 0, sizeof(fmap->signature)); |
| 94 | |
| 95 | boot_device_init(); |
| 96 | const struct region_device *boot_rdev = boot_device_ro(); |
| 97 | if (!boot_rdev) |
| 98 | return; |
| 99 | |
| 100 | /* memlayout statically guarantees that the FMAP_CACHE is big enough. */ |
| 101 | if (rdev_readat(boot_rdev, fmap, FMAP_OFFSET, FMAP_SIZE) != FMAP_SIZE) |
| 102 | return; |
Julius Werner | fdabf3f | 2020-05-06 17:06:35 -0700 | [diff] [blame] | 103 | if (verify_fmap(fmap)) |
Julius Werner | cefe89e | 2019-11-06 19:29:44 -0800 | [diff] [blame] | 104 | return; |
| 105 | report(fmap); |
| 106 | |
| 107 | register_cache: |
Julius Werner | c893197 | 2021-04-16 16:48:32 -0700 | [diff] [blame] | 108 | rdev_chain_mem(cache_rdev, fmap, FMAP_SIZE); |
Julius Werner | cefe89e | 2019-11-06 19:29:44 -0800 | [diff] [blame] | 109 | } |
| 110 | |
Furquan Shaikh | b33a2b0 | 2019-09-26 23:51:46 -0700 | [diff] [blame] | 111 | static int find_fmap_directory(struct region_device *fmrd) |
Aaron Durbin | 0424c95 | 2015-03-28 23:56:22 -0500 | [diff] [blame] | 112 | { |
| 113 | const struct region_device *boot; |
| 114 | struct fmap *fmap; |
Aaron Durbin | bf1e481 | 2016-05-10 15:12:08 -0500 | [diff] [blame] | 115 | size_t offset = FMAP_OFFSET; |
Aaron Durbin | 0424c95 | 2015-03-28 23:56:22 -0500 | [diff] [blame] | 116 | |
Julius Werner | cefe89e | 2019-11-06 19:29:44 -0800 | [diff] [blame] | 117 | /* Try FMAP cache first */ |
Julius Werner | c893197 | 2021-04-16 16:48:32 -0700 | [diff] [blame] | 118 | if (!region_device_sz(&fmap_cache)) |
Arthur Heymans | dba22d2 | 2019-11-20 19:57:49 +0100 | [diff] [blame] | 119 | setup_preram_cache(&fmap_cache); |
Julius Werner | c893197 | 2021-04-16 16:48:32 -0700 | [diff] [blame] | 120 | if (region_device_sz(&fmap_cache)) |
| 121 | return rdev_chain_full(fmrd, &fmap_cache); |
Patrick Rudolph | 6d787c2 | 2019-09-12 13:21:37 +0200 | [diff] [blame] | 122 | |
Julius Werner | ca71588 | 2023-11-02 16:20:17 -0700 | [diff] [blame] | 123 | /* Cache setup in pre-RAM stages can't fail, unless flash I/O in general failed. */ |
| 124 | if (!CONFIG(NO_FMAP_CACHE) && ENV_ROMSTAGE_OR_BEFORE) |
| 125 | return -1; |
| 126 | |
Aaron Durbin | 0424c95 | 2015-03-28 23:56:22 -0500 | [diff] [blame] | 127 | boot_device_init(); |
| 128 | boot = boot_device_ro(); |
| 129 | |
| 130 | if (boot == NULL) |
| 131 | return -1; |
| 132 | |
Julius Werner | c827c9b | 2023-11-09 12:20:59 -0800 | [diff] [blame] | 133 | fmap = rdev_mmap(boot, offset, |
| 134 | CONFIG(CBFS_VERIFICATION) ? FMAP_SIZE : sizeof(struct fmap)); |
Aaron Durbin | 0424c95 | 2015-03-28 23:56:22 -0500 | [diff] [blame] | 135 | |
| 136 | if (fmap == NULL) |
| 137 | return -1; |
| 138 | |
Julius Werner | fdabf3f | 2020-05-06 17:06:35 -0700 | [diff] [blame] | 139 | if (verify_fmap(fmap)) { |
Aaron Durbin | 0424c95 | 2015-03-28 23:56:22 -0500 | [diff] [blame] | 140 | rdev_munmap(boot, fmap); |
| 141 | return -1; |
| 142 | } |
| 143 | |
Julius Werner | cefe89e | 2019-11-06 19:29:44 -0800 | [diff] [blame] | 144 | report(fmap); |
Aaron Durbin | 0424c95 | 2015-03-28 23:56:22 -0500 | [diff] [blame] | 145 | |
| 146 | rdev_munmap(boot, fmap); |
| 147 | |
Julius Werner | cefe89e | 2019-11-06 19:29:44 -0800 | [diff] [blame] | 148 | return rdev_chain(fmrd, boot, offset, FMAP_SIZE); |
Aaron Durbin | 0424c95 | 2015-03-28 23:56:22 -0500 | [diff] [blame] | 149 | } |
| 150 | |
| 151 | int fmap_locate_area_as_rdev(const char *name, struct region_device *area) |
| 152 | { |
| 153 | struct region ar; |
| 154 | |
| 155 | if (fmap_locate_area(name, &ar)) |
| 156 | return -1; |
| 157 | |
| 158 | return boot_device_ro_subregion(&ar, area); |
| 159 | } |
| 160 | |
Aaron Durbin | bccaab8 | 2016-08-12 12:42:04 -0500 | [diff] [blame] | 161 | int fmap_locate_area_as_rdev_rw(const char *name, struct region_device *area) |
| 162 | { |
| 163 | struct region ar; |
| 164 | |
| 165 | if (fmap_locate_area(name, &ar)) |
| 166 | return -1; |
| 167 | |
| 168 | return boot_device_rw_subregion(&ar, area); |
| 169 | } |
| 170 | |
Aaron Durbin | 0424c95 | 2015-03-28 23:56:22 -0500 | [diff] [blame] | 171 | int fmap_locate_area(const char *name, struct region *ar) |
| 172 | { |
| 173 | struct region_device fmrd; |
| 174 | size_t offset; |
| 175 | |
Jakub Czapiga | 5446bdb | 2020-12-10 12:21:52 +0100 | [diff] [blame] | 176 | if (name == NULL || ar == NULL) |
| 177 | return -1; |
| 178 | |
Aaron Durbin | 0424c95 | 2015-03-28 23:56:22 -0500 | [diff] [blame] | 179 | if (find_fmap_directory(&fmrd)) |
| 180 | return -1; |
| 181 | |
| 182 | /* Start reading the areas just after fmap header. */ |
| 183 | offset = sizeof(struct fmap); |
| 184 | |
| 185 | while (1) { |
| 186 | struct fmap_area *area; |
| 187 | |
| 188 | area = rdev_mmap(&fmrd, offset, sizeof(*area)); |
| 189 | |
| 190 | if (area == NULL) |
| 191 | return -1; |
| 192 | |
| 193 | if (strcmp((const char *)area->name, name)) { |
| 194 | rdev_munmap(&fmrd, area); |
| 195 | offset += sizeof(struct fmap_area); |
| 196 | continue; |
| 197 | } |
| 198 | |
Duncan Laurie | bc2c0a3 | 2016-02-09 09:17:56 -0800 | [diff] [blame] | 199 | printk(BIOS_DEBUG, "FMAP: area %s found @ %x (%d bytes)\n", |
Krystian Hebel | 93f6b8a | 2020-09-30 18:23:14 +0200 | [diff] [blame] | 200 | name, le32toh(area->offset), le32toh(area->size)); |
Aaron Durbin | 0424c95 | 2015-03-28 23:56:22 -0500 | [diff] [blame] | 201 | |
Krystian Hebel | 93f6b8a | 2020-09-30 18:23:14 +0200 | [diff] [blame] | 202 | ar->offset = le32toh(area->offset); |
| 203 | ar->size = le32toh(area->size); |
Aaron Durbin | 0424c95 | 2015-03-28 23:56:22 -0500 | [diff] [blame] | 204 | |
| 205 | rdev_munmap(&fmrd, area); |
| 206 | |
| 207 | return 0; |
| 208 | } |
| 209 | |
| 210 | printk(BIOS_DEBUG, "FMAP: area %s not found\n", name); |
| 211 | |
| 212 | return -1; |
| 213 | } |
Patrick Georgi | 9952690 | 2015-07-09 11:27:44 +0200 | [diff] [blame] | 214 | |
| 215 | int fmap_find_region_name(const struct region * const ar, |
| 216 | char name[FMAP_STRLEN]) |
| 217 | { |
| 218 | struct region_device fmrd; |
| 219 | size_t offset; |
| 220 | |
Jakub Czapiga | 5446bdb | 2020-12-10 12:21:52 +0100 | [diff] [blame] | 221 | if (name == NULL || ar == NULL) |
| 222 | return -1; |
| 223 | |
Patrick Georgi | 9952690 | 2015-07-09 11:27:44 +0200 | [diff] [blame] | 224 | if (find_fmap_directory(&fmrd)) |
| 225 | return -1; |
| 226 | |
| 227 | /* Start reading the areas just after fmap header. */ |
| 228 | offset = sizeof(struct fmap); |
| 229 | |
| 230 | while (1) { |
| 231 | struct fmap_area *area; |
| 232 | |
| 233 | area = rdev_mmap(&fmrd, offset, sizeof(*area)); |
| 234 | |
| 235 | if (area == NULL) |
| 236 | return -1; |
| 237 | |
Nico Huber | d7612e9 | 2024-01-11 18:50:50 +0100 | [diff] [blame] | 238 | if (region_offset(ar) != le32toh(area->offset) || |
| 239 | region_sz(ar) != le32toh(area->size)) { |
Patrick Georgi | 9952690 | 2015-07-09 11:27:44 +0200 | [diff] [blame] | 240 | rdev_munmap(&fmrd, area); |
| 241 | offset += sizeof(struct fmap_area); |
| 242 | continue; |
| 243 | } |
| 244 | |
| 245 | printk(BIOS_DEBUG, "FMAP: area (%zx, %zx) found, named %s\n", |
Nico Huber | d7612e9 | 2024-01-11 18:50:50 +0100 | [diff] [blame] | 246 | region_offset(ar), region_sz(ar), area->name); |
Patrick Georgi | 9952690 | 2015-07-09 11:27:44 +0200 | [diff] [blame] | 247 | |
| 248 | memcpy(name, area->name, FMAP_STRLEN); |
| 249 | |
| 250 | rdev_munmap(&fmrd, area); |
| 251 | |
| 252 | return 0; |
| 253 | } |
| 254 | |
| 255 | printk(BIOS_DEBUG, "FMAP: area (%zx, %zx) not found\n", |
Nico Huber | d7612e9 | 2024-01-11 18:50:50 +0100 | [diff] [blame] | 256 | region_offset(ar), region_sz(ar)); |
Patrick Georgi | 9952690 | 2015-07-09 11:27:44 +0200 | [diff] [blame] | 257 | |
| 258 | return -1; |
| 259 | } |
T Michael Turney | 19fcc89 | 2019-03-20 14:37:34 -0700 | [diff] [blame] | 260 | |
| 261 | ssize_t fmap_read_area(const char *name, void *buffer, size_t size) |
| 262 | { |
| 263 | struct region_device rdev; |
| 264 | if (fmap_locate_area_as_rdev(name, &rdev)) |
| 265 | return -1; |
| 266 | return rdev_readat(&rdev, buffer, 0, |
| 267 | MIN(size, region_device_sz(&rdev))); |
| 268 | } |
| 269 | |
| 270 | ssize_t fmap_overwrite_area(const char *name, const void *buffer, size_t size) |
| 271 | { |
| 272 | struct region_device rdev; |
| 273 | |
| 274 | if (fmap_locate_area_as_rdev_rw(name, &rdev)) |
| 275 | return -1; |
| 276 | if (size > region_device_sz(&rdev)) |
| 277 | return -1; |
| 278 | if (rdev_eraseat(&rdev, 0, region_device_sz(&rdev)) < 0) |
| 279 | return -1; |
| 280 | return rdev_writeat(&rdev, buffer, 0, size); |
| 281 | } |
Patrick Rudolph | 6d787c2 | 2019-09-12 13:21:37 +0200 | [diff] [blame] | 282 | |
Kyösti Mälkki | 845f232 | 2022-04-06 10:53:17 +0300 | [diff] [blame] | 283 | static void fmap_register_cbmem_cache(void) |
Patrick Rudolph | 6d787c2 | 2019-09-12 13:21:37 +0200 | [diff] [blame] | 284 | { |
| 285 | const struct cbmem_entry *e; |
Patrick Rudolph | 6d787c2 | 2019-09-12 13:21:37 +0200 | [diff] [blame] | 286 | |
| 287 | /* Find the FMAP cache installed by previous stage */ |
| 288 | e = cbmem_entry_find(CBMEM_ID_FMAP); |
| 289 | /* Don't set fmap_cache so that find_fmap_directory will use regular path */ |
| 290 | if (!e) |
| 291 | return; |
| 292 | |
Julius Werner | c893197 | 2021-04-16 16:48:32 -0700 | [diff] [blame] | 293 | rdev_chain_mem(&fmap_cache, cbmem_entry_start(e), cbmem_entry_size(e)); |
Patrick Rudolph | 6d787c2 | 2019-09-12 13:21:37 +0200 | [diff] [blame] | 294 | } |
| 295 | |
| 296 | /* |
| 297 | * The main reason to copy the FMAP into CBMEM is to make it available to the |
| 298 | * OS on every architecture. As side effect use the CBMEM copy as cache. |
| 299 | */ |
Kyösti Mälkki | 845f232 | 2022-04-06 10:53:17 +0300 | [diff] [blame] | 300 | static void fmap_add_cbmem_cache(void) |
Patrick Rudolph | 6d787c2 | 2019-09-12 13:21:37 +0200 | [diff] [blame] | 301 | { |
| 302 | struct region_device fmrd; |
| 303 | |
| 304 | if (find_fmap_directory(&fmrd)) |
| 305 | return; |
| 306 | |
| 307 | /* Reloads the FMAP even on ACPI S3 resume */ |
| 308 | const size_t s = region_device_sz(&fmrd); |
| 309 | struct fmap *fmap = cbmem_add(CBMEM_ID_FMAP, s); |
| 310 | if (!fmap) { |
Julius Werner | e966595 | 2022-01-21 17:06:20 -0800 | [diff] [blame] | 311 | printk(BIOS_ERR, "Failed to allocate CBMEM\n"); |
Patrick Rudolph | 6d787c2 | 2019-09-12 13:21:37 +0200 | [diff] [blame] | 312 | return; |
| 313 | } |
| 314 | |
| 315 | const ssize_t ret = rdev_readat(&fmrd, fmap, 0, s); |
| 316 | if (ret != s) { |
Julius Werner | e966595 | 2022-01-21 17:06:20 -0800 | [diff] [blame] | 317 | printk(BIOS_ERR, "Failed to read FMAP into CBMEM\n"); |
Patrick Rudolph | 6d787c2 | 2019-09-12 13:21:37 +0200 | [diff] [blame] | 318 | cbmem_entry_remove(cbmem_entry_find(CBMEM_ID_FMAP)); |
| 319 | return; |
| 320 | } |
Kyösti Mälkki | 845f232 | 2022-04-06 10:53:17 +0300 | [diff] [blame] | 321 | } |
| 322 | |
| 323 | static void fmap_setup_cbmem_cache(int unused) |
| 324 | { |
Kyösti Mälkki | fa3bc04 | 2022-03-31 07:40:10 +0300 | [diff] [blame] | 325 | if (ENV_CREATES_CBMEM) |
Kyösti Mälkki | 845f232 | 2022-04-06 10:53:17 +0300 | [diff] [blame] | 326 | fmap_add_cbmem_cache(); |
Patrick Rudolph | 6d787c2 | 2019-09-12 13:21:37 +0200 | [diff] [blame] | 327 | |
| 328 | /* Finally advertise the cache for the current stage */ |
Kyösti Mälkki | 845f232 | 2022-04-06 10:53:17 +0300 | [diff] [blame] | 329 | fmap_register_cbmem_cache(); |
Patrick Rudolph | 6d787c2 | 2019-09-12 13:21:37 +0200 | [diff] [blame] | 330 | } |
| 331 | |
Kyösti Mälkki | fa3bc04 | 2022-03-31 07:40:10 +0300 | [diff] [blame] | 332 | CBMEM_READY_HOOK(fmap_setup_cbmem_cache); |