2009-03-11 H.J. Lu <hongjiu.lu@intel.com>
[binutils.git] / bfd / elf32-spu.c
blob115fa49c570579b9c6dec7e29196cecb45763447
1 /* SPU specific support for 32-bit ELF
3 Copyright 2006, 2007, 2008, 2009 Free Software Foundation, Inc.
5 This file is part of BFD, the Binary File Descriptor library.
7 This program is free software; you can redistribute it and/or modify
8 it under the terms of the GNU General Public License as published by
9 the Free Software Foundation; either version 3 of the License, or
10 (at your option) any later version.
12 This program is distributed in the hope that it will be useful,
13 but WITHOUT ANY WARRANTY; without even the implied warranty of
14 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
15 GNU General Public License for more details.
17 You should have received a copy of the GNU General Public License along
18 with this program; if not, write to the Free Software Foundation, Inc.,
19 51 Franklin Street - Fifth Floor, Boston, MA 02110-1301, USA. */
21 #include "sysdep.h"
22 #include "libiberty.h"
23 #include "bfd.h"
24 #include "bfdlink.h"
25 #include "libbfd.h"
26 #include "elf-bfd.h"
27 #include "elf/spu.h"
28 #include "elf32-spu.h"
30 /* We use RELA style relocs. Don't define USE_REL. */
32 static bfd_reloc_status_type spu_elf_rel9 (bfd *, arelent *, asymbol *,
33 void *, asection *,
34 bfd *, char **);
36 /* Values of type 'enum elf_spu_reloc_type' are used to index this
37 array, so it must be declared in the order of that type. */
39 static reloc_howto_type elf_howto_table[] = {
40 HOWTO (R_SPU_NONE, 0, 0, 0, FALSE, 0, complain_overflow_dont,
41 bfd_elf_generic_reloc, "SPU_NONE",
42 FALSE, 0, 0x00000000, FALSE),
43 HOWTO (R_SPU_ADDR10, 4, 2, 10, FALSE, 14, complain_overflow_bitfield,
44 bfd_elf_generic_reloc, "SPU_ADDR10",
45 FALSE, 0, 0x00ffc000, FALSE),
46 HOWTO (R_SPU_ADDR16, 2, 2, 16, FALSE, 7, complain_overflow_bitfield,
47 bfd_elf_generic_reloc, "SPU_ADDR16",
48 FALSE, 0, 0x007fff80, FALSE),
49 HOWTO (R_SPU_ADDR16_HI, 16, 2, 16, FALSE, 7, complain_overflow_bitfield,
50 bfd_elf_generic_reloc, "SPU_ADDR16_HI",
51 FALSE, 0, 0x007fff80, FALSE),
52 HOWTO (R_SPU_ADDR16_LO, 0, 2, 16, FALSE, 7, complain_overflow_dont,
53 bfd_elf_generic_reloc, "SPU_ADDR16_LO",
54 FALSE, 0, 0x007fff80, FALSE),
55 HOWTO (R_SPU_ADDR18, 0, 2, 18, FALSE, 7, complain_overflow_bitfield,
56 bfd_elf_generic_reloc, "SPU_ADDR18",
57 FALSE, 0, 0x01ffff80, FALSE),
58 HOWTO (R_SPU_ADDR32, 0, 2, 32, FALSE, 0, complain_overflow_dont,
59 bfd_elf_generic_reloc, "SPU_ADDR32",
60 FALSE, 0, 0xffffffff, FALSE),
61 HOWTO (R_SPU_REL16, 2, 2, 16, TRUE, 7, complain_overflow_bitfield,
62 bfd_elf_generic_reloc, "SPU_REL16",
63 FALSE, 0, 0x007fff80, TRUE),
64 HOWTO (R_SPU_ADDR7, 0, 2, 7, FALSE, 14, complain_overflow_dont,
65 bfd_elf_generic_reloc, "SPU_ADDR7",
66 FALSE, 0, 0x001fc000, FALSE),
67 HOWTO (R_SPU_REL9, 2, 2, 9, TRUE, 0, complain_overflow_signed,
68 spu_elf_rel9, "SPU_REL9",
69 FALSE, 0, 0x0180007f, TRUE),
70 HOWTO (R_SPU_REL9I, 2, 2, 9, TRUE, 0, complain_overflow_signed,
71 spu_elf_rel9, "SPU_REL9I",
72 FALSE, 0, 0x0000c07f, TRUE),
73 HOWTO (R_SPU_ADDR10I, 0, 2, 10, FALSE, 14, complain_overflow_signed,
74 bfd_elf_generic_reloc, "SPU_ADDR10I",
75 FALSE, 0, 0x00ffc000, FALSE),
76 HOWTO (R_SPU_ADDR16I, 0, 2, 16, FALSE, 7, complain_overflow_signed,
77 bfd_elf_generic_reloc, "SPU_ADDR16I",
78 FALSE, 0, 0x007fff80, FALSE),
79 HOWTO (R_SPU_REL32, 0, 2, 32, TRUE, 0, complain_overflow_dont,
80 bfd_elf_generic_reloc, "SPU_REL32",
81 FALSE, 0, 0xffffffff, TRUE),
82 HOWTO (R_SPU_ADDR16X, 0, 2, 16, FALSE, 7, complain_overflow_bitfield,
83 bfd_elf_generic_reloc, "SPU_ADDR16X",
84 FALSE, 0, 0x007fff80, FALSE),
85 HOWTO (R_SPU_PPU32, 0, 2, 32, FALSE, 0, complain_overflow_dont,
86 bfd_elf_generic_reloc, "SPU_PPU32",
87 FALSE, 0, 0xffffffff, FALSE),
88 HOWTO (R_SPU_PPU64, 0, 4, 64, FALSE, 0, complain_overflow_dont,
89 bfd_elf_generic_reloc, "SPU_PPU64",
90 FALSE, 0, -1, FALSE),
93 static struct bfd_elf_special_section const spu_elf_special_sections[] = {
94 { "._ea", 4, 0, SHT_PROGBITS, SHF_WRITE },
95 { ".toe", 4, 0, SHT_NOBITS, SHF_ALLOC },
96 { NULL, 0, 0, 0, 0 }
99 static enum elf_spu_reloc_type
100 spu_elf_bfd_to_reloc_type (bfd_reloc_code_real_type code)
102 switch (code)
104 default:
105 return R_SPU_NONE;
106 case BFD_RELOC_SPU_IMM10W:
107 return R_SPU_ADDR10;
108 case BFD_RELOC_SPU_IMM16W:
109 return R_SPU_ADDR16;
110 case BFD_RELOC_SPU_LO16:
111 return R_SPU_ADDR16_LO;
112 case BFD_RELOC_SPU_HI16:
113 return R_SPU_ADDR16_HI;
114 case BFD_RELOC_SPU_IMM18:
115 return R_SPU_ADDR18;
116 case BFD_RELOC_SPU_PCREL16:
117 return R_SPU_REL16;
118 case BFD_RELOC_SPU_IMM7:
119 return R_SPU_ADDR7;
120 case BFD_RELOC_SPU_IMM8:
121 return R_SPU_NONE;
122 case BFD_RELOC_SPU_PCREL9a:
123 return R_SPU_REL9;
124 case BFD_RELOC_SPU_PCREL9b:
125 return R_SPU_REL9I;
126 case BFD_RELOC_SPU_IMM10:
127 return R_SPU_ADDR10I;
128 case BFD_RELOC_SPU_IMM16:
129 return R_SPU_ADDR16I;
130 case BFD_RELOC_32:
131 return R_SPU_ADDR32;
132 case BFD_RELOC_32_PCREL:
133 return R_SPU_REL32;
134 case BFD_RELOC_SPU_PPU32:
135 return R_SPU_PPU32;
136 case BFD_RELOC_SPU_PPU64:
137 return R_SPU_PPU64;
141 static void
142 spu_elf_info_to_howto (bfd *abfd ATTRIBUTE_UNUSED,
143 arelent *cache_ptr,
144 Elf_Internal_Rela *dst)
146 enum elf_spu_reloc_type r_type;
148 r_type = (enum elf_spu_reloc_type) ELF32_R_TYPE (dst->r_info);
149 BFD_ASSERT (r_type < R_SPU_max);
150 cache_ptr->howto = &elf_howto_table[(int) r_type];
153 static reloc_howto_type *
154 spu_elf_reloc_type_lookup (bfd *abfd ATTRIBUTE_UNUSED,
155 bfd_reloc_code_real_type code)
157 enum elf_spu_reloc_type r_type = spu_elf_bfd_to_reloc_type (code);
159 if (r_type == R_SPU_NONE)
160 return NULL;
162 return elf_howto_table + r_type;
165 static reloc_howto_type *
166 spu_elf_reloc_name_lookup (bfd *abfd ATTRIBUTE_UNUSED,
167 const char *r_name)
169 unsigned int i;
171 for (i = 0; i < sizeof (elf_howto_table) / sizeof (elf_howto_table[0]); i++)
172 if (elf_howto_table[i].name != NULL
173 && strcasecmp (elf_howto_table[i].name, r_name) == 0)
174 return &elf_howto_table[i];
176 return NULL;
179 /* Apply R_SPU_REL9 and R_SPU_REL9I relocs. */
181 static bfd_reloc_status_type
182 spu_elf_rel9 (bfd *abfd, arelent *reloc_entry, asymbol *symbol,
183 void *data, asection *input_section,
184 bfd *output_bfd, char **error_message)
186 bfd_size_type octets;
187 bfd_vma val;
188 long insn;
190 /* If this is a relocatable link (output_bfd test tells us), just
191 call the generic function. Any adjustment will be done at final
192 link time. */
193 if (output_bfd != NULL)
194 return bfd_elf_generic_reloc (abfd, reloc_entry, symbol, data,
195 input_section, output_bfd, error_message);
197 if (reloc_entry->address > bfd_get_section_limit (abfd, input_section))
198 return bfd_reloc_outofrange;
199 octets = reloc_entry->address * bfd_octets_per_byte (abfd);
201 /* Get symbol value. */
202 val = 0;
203 if (!bfd_is_com_section (symbol->section))
204 val = symbol->value;
205 if (symbol->section->output_section)
206 val += symbol->section->output_section->vma;
208 val += reloc_entry->addend;
210 /* Make it pc-relative. */
211 val -= input_section->output_section->vma + input_section->output_offset;
213 val >>= 2;
214 if (val + 256 >= 512)
215 return bfd_reloc_overflow;
217 insn = bfd_get_32 (abfd, (bfd_byte *) data + octets);
219 /* Move two high bits of value to REL9I and REL9 position.
220 The mask will take care of selecting the right field. */
221 val = (val & 0x7f) | ((val & 0x180) << 7) | ((val & 0x180) << 16);
222 insn &= ~reloc_entry->howto->dst_mask;
223 insn |= val & reloc_entry->howto->dst_mask;
224 bfd_put_32 (abfd, insn, (bfd_byte *) data + octets);
225 return bfd_reloc_ok;
228 static bfd_boolean
229 spu_elf_new_section_hook (bfd *abfd, asection *sec)
231 if (!sec->used_by_bfd)
233 struct _spu_elf_section_data *sdata;
235 sdata = bfd_zalloc (abfd, sizeof (*sdata));
236 if (sdata == NULL)
237 return FALSE;
238 sec->used_by_bfd = sdata;
241 return _bfd_elf_new_section_hook (abfd, sec);
244 /* Set up overlay info for executables. */
246 static bfd_boolean
247 spu_elf_object_p (bfd *abfd)
249 if ((abfd->flags & (EXEC_P | DYNAMIC)) != 0)
251 unsigned int i, num_ovl, num_buf;
252 Elf_Internal_Phdr *phdr = elf_tdata (abfd)->phdr;
253 Elf_Internal_Ehdr *ehdr = elf_elfheader (abfd);
254 Elf_Internal_Phdr *last_phdr = NULL;
256 for (num_buf = 0, num_ovl = 0, i = 0; i < ehdr->e_phnum; i++, phdr++)
257 if (phdr->p_type == PT_LOAD && (phdr->p_flags & PF_OVERLAY) != 0)
259 unsigned int j;
261 ++num_ovl;
262 if (last_phdr == NULL
263 || ((last_phdr->p_vaddr ^ phdr->p_vaddr) & 0x3ffff) != 0)
264 ++num_buf;
265 last_phdr = phdr;
266 for (j = 1; j < elf_numsections (abfd); j++)
268 Elf_Internal_Shdr *shdr = elf_elfsections (abfd)[j];
270 if (ELF_IS_SECTION_IN_SEGMENT_MEMORY (shdr, phdr))
272 asection *sec = shdr->bfd_section;
273 spu_elf_section_data (sec)->u.o.ovl_index = num_ovl;
274 spu_elf_section_data (sec)->u.o.ovl_buf = num_buf;
279 return TRUE;
282 /* Specially mark defined symbols named _EAR_* with BSF_KEEP so that
283 strip --strip-unneeded will not remove them. */
285 static void
286 spu_elf_backend_symbol_processing (bfd *abfd ATTRIBUTE_UNUSED, asymbol *sym)
288 if (sym->name != NULL
289 && sym->section != bfd_abs_section_ptr
290 && strncmp (sym->name, "_EAR_", 5) == 0)
291 sym->flags |= BSF_KEEP;
294 /* SPU ELF linker hash table. */
296 struct spu_link_hash_table
298 struct elf_link_hash_table elf;
300 struct spu_elf_params *params;
302 /* Shortcuts to overlay sections. */
303 asection *ovtab;
304 asection *init;
305 asection *toe;
306 asection **ovl_sec;
308 /* Count of stubs in each overlay section. */
309 unsigned int *stub_count;
311 /* The stub section for each overlay section. */
312 asection **stub_sec;
314 struct elf_link_hash_entry *ovly_load;
315 struct elf_link_hash_entry *ovly_return;
316 unsigned long ovly_load_r_symndx;
318 /* Number of overlay buffers. */
319 unsigned int num_buf;
321 /* Total number of overlays. */
322 unsigned int num_overlays;
324 /* For soft icache. */
325 unsigned int line_size_log2;
326 unsigned int num_lines_log2;
328 /* How much memory we have. */
329 unsigned int local_store;
330 /* Local store --auto-overlay should reserve for non-overlay
331 functions and data. */
332 unsigned int overlay_fixed;
333 /* Local store --auto-overlay should reserve for stack and heap. */
334 unsigned int reserved;
335 /* If reserved is not specified, stack analysis will calculate a value
336 for the stack. This parameter adjusts that value to allow for
337 negative sp access (the ABI says 2000 bytes below sp are valid,
338 and the overlay manager uses some of this area). */
339 int extra_stack_space;
340 /* Count of overlay stubs needed in non-overlay area. */
341 unsigned int non_ovly_stub;
343 /* Set on error. */
344 unsigned int stub_err : 1;
347 /* Hijack the generic got fields for overlay stub accounting. */
349 struct got_entry
351 struct got_entry *next;
352 unsigned int ovl;
353 union {
354 bfd_vma addend;
355 bfd_vma br_addr;
357 bfd_vma stub_addr;
360 #define spu_hash_table(p) \
361 ((struct spu_link_hash_table *) ((p)->hash))
363 struct call_info
365 struct function_info *fun;
366 struct call_info *next;
367 unsigned int count;
368 unsigned int max_depth;
369 unsigned int is_tail : 1;
370 unsigned int is_pasted : 1;
371 unsigned int priority : 13;
374 struct function_info
376 /* List of functions called. Also branches to hot/cold part of
377 function. */
378 struct call_info *call_list;
379 /* For hot/cold part of function, point to owner. */
380 struct function_info *start;
381 /* Symbol at start of function. */
382 union {
383 Elf_Internal_Sym *sym;
384 struct elf_link_hash_entry *h;
385 } u;
386 /* Function section. */
387 asection *sec;
388 asection *rodata;
389 /* Where last called from, and number of sections called from. */
390 asection *last_caller;
391 unsigned int call_count;
392 /* Address range of (this part of) function. */
393 bfd_vma lo, hi;
394 /* Offset where we found a store of lr, or -1 if none found. */
395 bfd_vma lr_store;
396 /* Offset where we found the stack adjustment insn. */
397 bfd_vma sp_adjust;
398 /* Stack usage. */
399 int stack;
400 /* Distance from root of call tree. Tail and hot/cold branches
401 count as one deeper. We aren't counting stack frames here. */
402 unsigned int depth;
403 /* Set if global symbol. */
404 unsigned int global : 1;
405 /* Set if known to be start of function (as distinct from a hunk
406 in hot/cold section. */
407 unsigned int is_func : 1;
408 /* Set if not a root node. */
409 unsigned int non_root : 1;
410 /* Flags used during call tree traversal. It's cheaper to replicate
411 the visit flags than have one which needs clearing after a traversal. */
412 unsigned int visit1 : 1;
413 unsigned int visit2 : 1;
414 unsigned int marking : 1;
415 unsigned int visit3 : 1;
416 unsigned int visit4 : 1;
417 unsigned int visit5 : 1;
418 unsigned int visit6 : 1;
419 unsigned int visit7 : 1;
422 struct spu_elf_stack_info
424 int num_fun;
425 int max_fun;
426 /* Variable size array describing functions, one per contiguous
427 address range belonging to a function. */
428 struct function_info fun[1];
431 static struct function_info *find_function (asection *, bfd_vma,
432 struct bfd_link_info *);
434 /* Create a spu ELF linker hash table. */
436 static struct bfd_link_hash_table *
437 spu_elf_link_hash_table_create (bfd *abfd)
439 struct spu_link_hash_table *htab;
441 htab = bfd_malloc (sizeof (*htab));
442 if (htab == NULL)
443 return NULL;
445 if (!_bfd_elf_link_hash_table_init (&htab->elf, abfd,
446 _bfd_elf_link_hash_newfunc,
447 sizeof (struct elf_link_hash_entry)))
449 free (htab);
450 return NULL;
453 memset (&htab->ovtab, 0,
454 sizeof (*htab) - offsetof (struct spu_link_hash_table, ovtab));
456 htab->elf.init_got_refcount.refcount = 0;
457 htab->elf.init_got_refcount.glist = NULL;
458 htab->elf.init_got_offset.offset = 0;
459 htab->elf.init_got_offset.glist = NULL;
460 return &htab->elf.root;
463 void
464 spu_elf_setup (struct bfd_link_info *info, struct spu_elf_params *params)
466 struct spu_link_hash_table *htab = spu_hash_table (info);
467 htab->params = params;
468 htab->line_size_log2 = bfd_log2 (htab->params->line_size);
469 htab->num_lines_log2 = bfd_log2 (htab->params->num_lines);
472 /* Find the symbol for the given R_SYMNDX in IBFD and set *HP and *SYMP
473 to (hash, NULL) for global symbols, and (NULL, sym) for locals. Set
474 *SYMSECP to the symbol's section. *LOCSYMSP caches local syms. */
476 static bfd_boolean
477 get_sym_h (struct elf_link_hash_entry **hp,
478 Elf_Internal_Sym **symp,
479 asection **symsecp,
480 Elf_Internal_Sym **locsymsp,
481 unsigned long r_symndx,
482 bfd *ibfd)
484 Elf_Internal_Shdr *symtab_hdr = &elf_tdata (ibfd)->symtab_hdr;
486 if (r_symndx >= symtab_hdr->sh_info)
488 struct elf_link_hash_entry **sym_hashes = elf_sym_hashes (ibfd);
489 struct elf_link_hash_entry *h;
491 h = sym_hashes[r_symndx - symtab_hdr->sh_info];
492 while (h->root.type == bfd_link_hash_indirect
493 || h->root.type == bfd_link_hash_warning)
494 h = (struct elf_link_hash_entry *) h->root.u.i.link;
496 if (hp != NULL)
497 *hp = h;
499 if (symp != NULL)
500 *symp = NULL;
502 if (symsecp != NULL)
504 asection *symsec = NULL;
505 if (h->root.type == bfd_link_hash_defined
506 || h->root.type == bfd_link_hash_defweak)
507 symsec = h->root.u.def.section;
508 *symsecp = symsec;
511 else
513 Elf_Internal_Sym *sym;
514 Elf_Internal_Sym *locsyms = *locsymsp;
516 if (locsyms == NULL)
518 locsyms = (Elf_Internal_Sym *) symtab_hdr->contents;
519 if (locsyms == NULL)
520 locsyms = bfd_elf_get_elf_syms (ibfd, symtab_hdr,
521 symtab_hdr->sh_info,
522 0, NULL, NULL, NULL);
523 if (locsyms == NULL)
524 return FALSE;
525 *locsymsp = locsyms;
527 sym = locsyms + r_symndx;
529 if (hp != NULL)
530 *hp = NULL;
532 if (symp != NULL)
533 *symp = sym;
535 if (symsecp != NULL)
536 *symsecp = bfd_section_from_elf_index (ibfd, sym->st_shndx);
539 return TRUE;
542 /* Create the note section if not already present. This is done early so
543 that the linker maps the sections to the right place in the output. */
545 bfd_boolean
546 spu_elf_create_sections (struct bfd_link_info *info)
548 bfd *ibfd;
550 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
551 if (bfd_get_section_by_name (ibfd, SPU_PTNOTE_SPUNAME) != NULL)
552 break;
554 if (ibfd == NULL)
556 /* Make SPU_PTNOTE_SPUNAME section. */
557 asection *s;
558 size_t name_len;
559 size_t size;
560 bfd_byte *data;
561 flagword flags;
563 ibfd = info->input_bfds;
564 flags = SEC_LOAD | SEC_READONLY | SEC_HAS_CONTENTS | SEC_IN_MEMORY;
565 s = bfd_make_section_anyway_with_flags (ibfd, SPU_PTNOTE_SPUNAME, flags);
566 if (s == NULL
567 || !bfd_set_section_alignment (ibfd, s, 4))
568 return FALSE;
570 name_len = strlen (bfd_get_filename (info->output_bfd)) + 1;
571 size = 12 + ((sizeof (SPU_PLUGIN_NAME) + 3) & -4);
572 size += (name_len + 3) & -4;
574 if (!bfd_set_section_size (ibfd, s, size))
575 return FALSE;
577 data = bfd_zalloc (ibfd, size);
578 if (data == NULL)
579 return FALSE;
581 bfd_put_32 (ibfd, sizeof (SPU_PLUGIN_NAME), data + 0);
582 bfd_put_32 (ibfd, name_len, data + 4);
583 bfd_put_32 (ibfd, 1, data + 8);
584 memcpy (data + 12, SPU_PLUGIN_NAME, sizeof (SPU_PLUGIN_NAME));
585 memcpy (data + 12 + ((sizeof (SPU_PLUGIN_NAME) + 3) & -4),
586 bfd_get_filename (info->output_bfd), name_len);
587 s->contents = data;
590 return TRUE;
593 /* qsort predicate to sort sections by vma. */
595 static int
596 sort_sections (const void *a, const void *b)
598 const asection *const *s1 = a;
599 const asection *const *s2 = b;
600 bfd_signed_vma delta = (*s1)->vma - (*s2)->vma;
602 if (delta != 0)
603 return delta < 0 ? -1 : 1;
605 return (*s1)->index - (*s2)->index;
608 /* Identify overlays in the output bfd, and number them. */
610 bfd_boolean
611 spu_elf_find_overlays (struct bfd_link_info *info)
613 struct spu_link_hash_table *htab = spu_hash_table (info);
614 asection **alloc_sec;
615 unsigned int i, n, ovl_index, num_buf;
616 asection *s;
617 bfd_vma ovl_end;
618 const char *ovly_mgr_entry;
620 if (info->output_bfd->section_count < 2)
621 return FALSE;
623 alloc_sec
624 = bfd_malloc (info->output_bfd->section_count * sizeof (*alloc_sec));
625 if (alloc_sec == NULL)
626 return FALSE;
628 /* Pick out all the alloced sections. */
629 for (n = 0, s = info->output_bfd->sections; s != NULL; s = s->next)
630 if ((s->flags & SEC_ALLOC) != 0
631 && (s->flags & (SEC_LOAD | SEC_THREAD_LOCAL)) != SEC_THREAD_LOCAL
632 && s->size != 0)
633 alloc_sec[n++] = s;
635 if (n == 0)
637 free (alloc_sec);
638 return FALSE;
641 /* Sort them by vma. */
642 qsort (alloc_sec, n, sizeof (*alloc_sec), sort_sections);
644 ovl_end = alloc_sec[0]->vma + alloc_sec[0]->size;
645 if (htab->params->ovly_flavour == ovly_soft_icache)
647 /* Look for an overlapping vma to find the first overlay section. */
648 bfd_vma vma_start = 0;
649 bfd_vma lma_start = 0;
651 for (i = 1; i < n; i++)
653 s = alloc_sec[i];
654 if (s->vma < ovl_end)
656 asection *s0 = alloc_sec[i - 1];
657 vma_start = s0->vma;
658 lma_start = s0->lma;
659 ovl_end = (s0->vma
660 + ((bfd_vma) 1
661 << (htab->num_lines_log2 + htab->line_size_log2)));
662 --i;
663 break;
665 else
666 ovl_end = s->vma + s->size;
669 /* Now find any sections within the cache area. */
670 for (ovl_index = 0, num_buf = 0; i < n; i++)
672 s = alloc_sec[i];
673 if (s->vma >= ovl_end)
674 break;
676 /* A section in an overlay area called .ovl.init is not
677 an overlay, in the sense that it might be loaded in
678 by the overlay manager, but rather the initial
679 section contents for the overlay buffer. */
680 if (strncmp (s->name, ".ovl.init", 9) != 0)
682 num_buf = ((s->vma - vma_start) >> htab->line_size_log2) + 1;
683 if (((s->vma - vma_start) & (htab->params->line_size - 1))
684 || ((s->lma - lma_start) & (htab->params->line_size - 1)))
686 info->callbacks->einfo (_("%X%P: overlay section %A "
687 "does not start on a cache line.\n"),
689 bfd_set_error (bfd_error_bad_value);
690 return FALSE;
692 else if (s->size > htab->params->line_size)
694 info->callbacks->einfo (_("%X%P: overlay section %A "
695 "is larger than a cache line.\n"),
697 bfd_set_error (bfd_error_bad_value);
698 return FALSE;
701 alloc_sec[ovl_index++] = s;
702 spu_elf_section_data (s)->u.o.ovl_index
703 = ((s->lma - lma_start) >> htab->line_size_log2) + 1;
704 spu_elf_section_data (s)->u.o.ovl_buf = num_buf;
708 /* Ensure there are no more overlay sections. */
709 for ( ; i < n; i++)
711 s = alloc_sec[i];
712 if (s->vma < ovl_end)
714 info->callbacks->einfo (_("%X%P: overlay section %A "
715 "is not in cache area.\n"),
716 alloc_sec[i-1]);
717 bfd_set_error (bfd_error_bad_value);
718 return FALSE;
720 else
721 ovl_end = s->vma + s->size;
724 else
726 /* Look for overlapping vmas. Any with overlap must be overlays.
727 Count them. Also count the number of overlay regions. */
728 for (ovl_index = 0, num_buf = 0, i = 1; i < n; i++)
730 s = alloc_sec[i];
731 if (s->vma < ovl_end)
733 asection *s0 = alloc_sec[i - 1];
735 if (spu_elf_section_data (s0)->u.o.ovl_index == 0)
737 ++num_buf;
738 if (strncmp (s0->name, ".ovl.init", 9) != 0)
740 alloc_sec[ovl_index] = s0;
741 spu_elf_section_data (s0)->u.o.ovl_index = ++ovl_index;
742 spu_elf_section_data (s0)->u.o.ovl_buf = num_buf;
744 else
745 ovl_end = s->vma + s->size;
747 if (strncmp (s->name, ".ovl.init", 9) != 0)
749 alloc_sec[ovl_index] = s;
750 spu_elf_section_data (s)->u.o.ovl_index = ++ovl_index;
751 spu_elf_section_data (s)->u.o.ovl_buf = num_buf;
752 if (s0->vma != s->vma)
754 info->callbacks->einfo (_("%X%P: overlay sections %A "
755 "and %A do not start at the "
756 "same address.\n"),
757 s0, s);
758 bfd_set_error (bfd_error_bad_value);
759 return FALSE;
761 if (ovl_end < s->vma + s->size)
762 ovl_end = s->vma + s->size;
765 else
766 ovl_end = s->vma + s->size;
770 htab->num_overlays = ovl_index;
771 htab->num_buf = num_buf;
772 htab->ovl_sec = alloc_sec;
773 ovly_mgr_entry = "__ovly_load";
774 if (htab->params->ovly_flavour == ovly_soft_icache)
775 ovly_mgr_entry = "__icache_br_handler";
776 htab->ovly_load = elf_link_hash_lookup (&htab->elf, ovly_mgr_entry,
777 FALSE, FALSE, FALSE);
778 if (htab->params->ovly_flavour != ovly_soft_icache)
779 htab->ovly_return = elf_link_hash_lookup (&htab->elf, "__ovly_return",
780 FALSE, FALSE, FALSE);
781 return ovl_index != 0;
784 /* Non-zero to use bra in overlay stubs rather than br. */
785 #define BRA_STUBS 0
787 #define BRA 0x30000000
788 #define BRASL 0x31000000
789 #define BR 0x32000000
790 #define BRSL 0x33000000
791 #define NOP 0x40200000
792 #define LNOP 0x00200000
793 #define ILA 0x42000000
795 /* Return true for all relative and absolute branch instructions.
796 bra 00110000 0..
797 brasl 00110001 0..
798 br 00110010 0..
799 brsl 00110011 0..
800 brz 00100000 0..
801 brnz 00100001 0..
802 brhz 00100010 0..
803 brhnz 00100011 0.. */
805 static bfd_boolean
806 is_branch (const unsigned char *insn)
808 return (insn[0] & 0xec) == 0x20 && (insn[1] & 0x80) == 0;
811 /* Return true for all indirect branch instructions.
812 bi 00110101 000
813 bisl 00110101 001
814 iret 00110101 010
815 bisled 00110101 011
816 biz 00100101 000
817 binz 00100101 001
818 bihz 00100101 010
819 bihnz 00100101 011 */
821 static bfd_boolean
822 is_indirect_branch (const unsigned char *insn)
824 return (insn[0] & 0xef) == 0x25 && (insn[1] & 0x80) == 0;
827 /* Return true for branch hint instructions.
828 hbra 0001000..
829 hbrr 0001001.. */
831 static bfd_boolean
832 is_hint (const unsigned char *insn)
834 return (insn[0] & 0xfc) == 0x10;
837 /* True if INPUT_SECTION might need overlay stubs. */
839 static bfd_boolean
840 maybe_needs_stubs (asection *input_section)
842 /* No stubs for debug sections and suchlike. */
843 if ((input_section->flags & SEC_ALLOC) == 0)
844 return FALSE;
846 /* No stubs for link-once sections that will be discarded. */
847 if (input_section->output_section == bfd_abs_section_ptr)
848 return FALSE;
850 /* Don't create stubs for .eh_frame references. */
851 if (strcmp (input_section->name, ".eh_frame") == 0)
852 return FALSE;
854 return TRUE;
857 enum _stub_type
859 no_stub,
860 call_ovl_stub,
861 br000_ovl_stub,
862 br001_ovl_stub,
863 br010_ovl_stub,
864 br011_ovl_stub,
865 br100_ovl_stub,
866 br101_ovl_stub,
867 br110_ovl_stub,
868 br111_ovl_stub,
869 nonovl_stub,
870 stub_error
873 /* Return non-zero if this reloc symbol should go via an overlay stub.
874 Return 2 if the stub must be in non-overlay area. */
876 static enum _stub_type
877 needs_ovl_stub (struct elf_link_hash_entry *h,
878 Elf_Internal_Sym *sym,
879 asection *sym_sec,
880 asection *input_section,
881 Elf_Internal_Rela *irela,
882 bfd_byte *contents,
883 struct bfd_link_info *info)
885 struct spu_link_hash_table *htab = spu_hash_table (info);
886 enum elf_spu_reloc_type r_type;
887 unsigned int sym_type;
888 bfd_boolean branch, hint, call;
889 enum _stub_type ret = no_stub;
890 bfd_byte insn[4];
892 if (sym_sec == NULL
893 || sym_sec->output_section == bfd_abs_section_ptr
894 || spu_elf_section_data (sym_sec->output_section) == NULL)
895 return ret;
897 if (h != NULL)
899 /* Ensure no stubs for user supplied overlay manager syms. */
900 if (h == htab->ovly_load || h == htab->ovly_return)
901 return ret;
903 /* setjmp always goes via an overlay stub, because then the return
904 and hence the longjmp goes via __ovly_return. That magically
905 makes setjmp/longjmp between overlays work. */
906 if (strncmp (h->root.root.string, "setjmp", 6) == 0
907 && (h->root.root.string[6] == '\0' || h->root.root.string[6] == '@'))
908 ret = call_ovl_stub;
911 if (h != NULL)
912 sym_type = h->type;
913 else
914 sym_type = ELF_ST_TYPE (sym->st_info);
916 r_type = ELF32_R_TYPE (irela->r_info);
917 branch = FALSE;
918 hint = FALSE;
919 call = FALSE;
920 if (r_type == R_SPU_REL16 || r_type == R_SPU_ADDR16)
922 if (contents == NULL)
924 contents = insn;
925 if (!bfd_get_section_contents (input_section->owner,
926 input_section,
927 contents,
928 irela->r_offset, 4))
929 return stub_error;
931 else
932 contents += irela->r_offset;
934 branch = is_branch (contents);
935 hint = is_hint (contents);
936 if (branch || hint)
938 call = (contents[0] & 0xfd) == 0x31;
939 if (call
940 && sym_type != STT_FUNC
941 && contents != insn)
943 /* It's common for people to write assembly and forget
944 to give function symbols the right type. Handle
945 calls to such symbols, but warn so that (hopefully)
946 people will fix their code. We need the symbol
947 type to be correct to distinguish function pointer
948 initialisation from other pointer initialisations. */
949 const char *sym_name;
951 if (h != NULL)
952 sym_name = h->root.root.string;
953 else
955 Elf_Internal_Shdr *symtab_hdr;
956 symtab_hdr = &elf_tdata (input_section->owner)->symtab_hdr;
957 sym_name = bfd_elf_sym_name (input_section->owner,
958 symtab_hdr,
959 sym,
960 sym_sec);
962 (*_bfd_error_handler) (_("warning: call to non-function"
963 " symbol %s defined in %B"),
964 sym_sec->owner, sym_name);
970 if ((!branch && htab->params->ovly_flavour == ovly_soft_icache)
971 || (sym_type != STT_FUNC
972 && !(branch || hint)
973 && (sym_sec->flags & SEC_CODE) == 0))
974 return no_stub;
976 /* Usually, symbols in non-overlay sections don't need stubs. */
977 if (spu_elf_section_data (sym_sec->output_section)->u.o.ovl_index == 0
978 && !htab->params->non_overlay_stubs)
979 return ret;
981 /* A reference from some other section to a symbol in an overlay
982 section needs a stub. */
983 if (spu_elf_section_data (sym_sec->output_section)->u.o.ovl_index
984 != spu_elf_section_data (input_section->output_section)->u.o.ovl_index)
986 if (call || sym_type == STT_FUNC)
987 ret = call_ovl_stub;
988 else
990 ret = br000_ovl_stub;
992 if (branch)
994 unsigned int lrlive = (contents[1] & 0x70) >> 4;
995 ret += lrlive;
1000 /* If this insn isn't a branch then we are possibly taking the
1001 address of a function and passing it out somehow. Soft-icache code
1002 always generates inline code to do indirect branches. */
1003 if (!(branch || hint)
1004 && sym_type == STT_FUNC
1005 && htab->params->ovly_flavour != ovly_soft_icache)
1006 ret = nonovl_stub;
1008 return ret;
1011 static bfd_boolean
1012 count_stub (struct spu_link_hash_table *htab,
1013 bfd *ibfd,
1014 asection *isec,
1015 enum _stub_type stub_type,
1016 struct elf_link_hash_entry *h,
1017 const Elf_Internal_Rela *irela)
1019 unsigned int ovl = 0;
1020 struct got_entry *g, **head;
1021 bfd_vma addend;
1023 /* If this instruction is a branch or call, we need a stub
1024 for it. One stub per function per overlay.
1025 If it isn't a branch, then we are taking the address of
1026 this function so need a stub in the non-overlay area
1027 for it. One stub per function. */
1028 if (stub_type != nonovl_stub)
1029 ovl = spu_elf_section_data (isec->output_section)->u.o.ovl_index;
1031 if (h != NULL)
1032 head = &h->got.glist;
1033 else
1035 if (elf_local_got_ents (ibfd) == NULL)
1037 bfd_size_type amt = (elf_tdata (ibfd)->symtab_hdr.sh_info
1038 * sizeof (*elf_local_got_ents (ibfd)));
1039 elf_local_got_ents (ibfd) = bfd_zmalloc (amt);
1040 if (elf_local_got_ents (ibfd) == NULL)
1041 return FALSE;
1043 head = elf_local_got_ents (ibfd) + ELF32_R_SYM (irela->r_info);
1046 if (htab->params->ovly_flavour == ovly_soft_icache)
1048 htab->stub_count[ovl] += 1;
1049 return TRUE;
1052 addend = 0;
1053 if (irela != NULL)
1054 addend = irela->r_addend;
1056 if (ovl == 0)
1058 struct got_entry *gnext;
1060 for (g = *head; g != NULL; g = g->next)
1061 if (g->addend == addend && g->ovl == 0)
1062 break;
1064 if (g == NULL)
1066 /* Need a new non-overlay area stub. Zap other stubs. */
1067 for (g = *head; g != NULL; g = gnext)
1069 gnext = g->next;
1070 if (g->addend == addend)
1072 htab->stub_count[g->ovl] -= 1;
1073 free (g);
1078 else
1080 for (g = *head; g != NULL; g = g->next)
1081 if (g->addend == addend && (g->ovl == ovl || g->ovl == 0))
1082 break;
1085 if (g == NULL)
1087 g = bfd_malloc (sizeof *g);
1088 if (g == NULL)
1089 return FALSE;
1090 g->ovl = ovl;
1091 g->addend = addend;
1092 g->stub_addr = (bfd_vma) -1;
1093 g->next = *head;
1094 *head = g;
1096 htab->stub_count[ovl] += 1;
1099 return TRUE;
1102 /* Support two sizes of overlay stubs, a slower more compact stub of two
1103 intructions, and a faster stub of four instructions. */
1105 static unsigned int
1106 ovl_stub_size (enum _ovly_flavour ovly_flavour)
1108 return 8 << ovly_flavour;
1111 /* Two instruction overlay stubs look like:
1113 brsl $75,__ovly_load
1114 .word target_ovl_and_address
1116 ovl_and_address is a word with the overlay number in the top 14 bits
1117 and local store address in the bottom 18 bits.
1119 Four instruction overlay stubs look like:
1121 ila $78,ovl_number
1122 lnop
1123 ila $79,target_address
1124 br __ovly_load
1126 Software icache stubs are:
1128 .word target_index
1129 .word target_ia;
1130 .word lrlive_branchlocalstoreaddr;
1131 brasl $75,__icache_br_handler
1132 .quad xor_pattern
1135 static bfd_boolean
1136 build_stub (struct bfd_link_info *info,
1137 bfd *ibfd,
1138 asection *isec,
1139 enum _stub_type stub_type,
1140 struct elf_link_hash_entry *h,
1141 const Elf_Internal_Rela *irela,
1142 bfd_vma dest,
1143 asection *dest_sec)
1145 struct spu_link_hash_table *htab = spu_hash_table (info);
1146 unsigned int ovl, dest_ovl, set_id;
1147 struct got_entry *g, **head;
1148 asection *sec;
1149 bfd_vma addend, from, to, br_dest, patt;
1150 unsigned int lrlive;
1152 ovl = 0;
1153 if (stub_type != nonovl_stub)
1154 ovl = spu_elf_section_data (isec->output_section)->u.o.ovl_index;
1156 if (h != NULL)
1157 head = &h->got.glist;
1158 else
1159 head = elf_local_got_ents (ibfd) + ELF32_R_SYM (irela->r_info);
1161 addend = 0;
1162 if (irela != NULL)
1163 addend = irela->r_addend;
1165 if (htab->params->ovly_flavour == ovly_soft_icache)
1167 g = bfd_malloc (sizeof *g);
1168 if (g == NULL)
1169 return FALSE;
1170 g->ovl = ovl;
1171 g->br_addr = 0;
1172 if (irela != NULL)
1173 g->br_addr = (irela->r_offset
1174 + isec->output_offset
1175 + isec->output_section->vma);
1176 g->next = *head;
1177 *head = g;
1179 else
1181 for (g = *head; g != NULL; g = g->next)
1182 if (g->addend == addend && (g->ovl == ovl || g->ovl == 0))
1183 break;
1184 if (g == NULL)
1185 abort ();
1187 if (g->ovl == 0 && ovl != 0)
1188 return TRUE;
1190 if (g->stub_addr != (bfd_vma) -1)
1191 return TRUE;
1194 sec = htab->stub_sec[ovl];
1195 dest += dest_sec->output_offset + dest_sec->output_section->vma;
1196 from = sec->size + sec->output_offset + sec->output_section->vma;
1197 g->stub_addr = from;
1198 to = (htab->ovly_load->root.u.def.value
1199 + htab->ovly_load->root.u.def.section->output_offset
1200 + htab->ovly_load->root.u.def.section->output_section->vma);
1202 if (((dest | to | from) & 3) != 0)
1204 htab->stub_err = 1;
1205 return FALSE;
1207 dest_ovl = spu_elf_section_data (dest_sec->output_section)->u.o.ovl_index;
1209 switch (htab->params->ovly_flavour)
1211 case ovly_normal:
1212 bfd_put_32 (sec->owner, ILA + ((dest_ovl << 7) & 0x01ffff80) + 78,
1213 sec->contents + sec->size);
1214 bfd_put_32 (sec->owner, LNOP,
1215 sec->contents + sec->size + 4);
1216 bfd_put_32 (sec->owner, ILA + ((dest << 7) & 0x01ffff80) + 79,
1217 sec->contents + sec->size + 8);
1218 if (!BRA_STUBS)
1219 bfd_put_32 (sec->owner, BR + (((to - (from + 12)) << 5) & 0x007fff80),
1220 sec->contents + sec->size + 12);
1221 else
1222 bfd_put_32 (sec->owner, BRA + ((to << 5) & 0x007fff80),
1223 sec->contents + sec->size + 12);
1224 break;
1226 case ovly_compact:
1227 if (!BRA_STUBS)
1228 bfd_put_32 (sec->owner, BRSL + (((to - from) << 5) & 0x007fff80) + 75,
1229 sec->contents + sec->size);
1230 else
1231 bfd_put_32 (sec->owner, BRASL + ((to << 5) & 0x007fff80) + 75,
1232 sec->contents + sec->size);
1233 bfd_put_32 (sec->owner, (dest & 0x3ffff) | (dest_ovl << 18),
1234 sec->contents + sec->size + 4);
1235 break;
1237 case ovly_soft_icache:
1238 lrlive = 0;
1239 if (stub_type == nonovl_stub)
1241 else if (stub_type == call_ovl_stub)
1242 /* A brsl makes lr live and *(*sp+16) is live.
1243 Tail calls have the same liveness. */
1244 lrlive = 5;
1245 else if (!htab->params->lrlive_analysis)
1246 /* Assume stack frame and lr save. */
1247 lrlive = 1;
1248 else if (irela != NULL)
1250 /* Analyse branch instructions. */
1251 struct function_info *caller;
1252 bfd_vma off;
1254 caller = find_function (isec, irela->r_offset, info);
1255 if (caller->start == NULL)
1256 off = irela->r_offset;
1257 else
1259 struct function_info *found = NULL;
1261 /* Find the earliest piece of this function that
1262 has frame adjusting instructions. We might
1263 see dynamic frame adjustment (eg. for alloca)
1264 in some later piece, but functions using
1265 alloca always set up a frame earlier. Frame
1266 setup instructions are always in one piece. */
1267 if (caller->lr_store != (bfd_vma) -1
1268 || caller->sp_adjust != (bfd_vma) -1)
1269 found = caller;
1270 while (caller->start != NULL)
1272 caller = caller->start;
1273 if (caller->lr_store != (bfd_vma) -1
1274 || caller->sp_adjust != (bfd_vma) -1)
1275 found = caller;
1277 if (found != NULL)
1278 caller = found;
1279 off = (bfd_vma) -1;
1282 if (off > caller->sp_adjust)
1284 if (off > caller->lr_store)
1285 /* Only *(*sp+16) is live. */
1286 lrlive = 1;
1287 else
1288 /* If no lr save, then we must be in a
1289 leaf function with a frame.
1290 lr is still live. */
1291 lrlive = 4;
1293 else if (off > caller->lr_store)
1295 /* Between lr save and stack adjust. */
1296 lrlive = 3;
1297 /* This should never happen since prologues won't
1298 be split here. */
1299 BFD_ASSERT (0);
1301 else
1302 /* On entry to function. */
1303 lrlive = 5;
1305 if (stub_type != br000_ovl_stub
1306 && lrlive != stub_type - br000_ovl_stub)
1307 info->callbacks->einfo (_("%A:0x%v lrlive .brinfo (%u) differs "
1308 "from analysis (%u)\n"),
1309 isec, irela->r_offset, lrlive,
1310 stub_type - br000_ovl_stub);
1313 /* If given lrlive info via .brinfo, use it. */
1314 if (stub_type > br000_ovl_stub)
1315 lrlive = stub_type - br000_ovl_stub;
1317 /* The branch that uses this stub goes to stub_addr + 12. We'll
1318 set up an xor pattern that can be used by the icache manager
1319 to modify this branch to go directly to its destination. */
1320 g->stub_addr += 12;
1321 br_dest = g->stub_addr;
1322 if (irela == NULL)
1324 /* Except in the case of _SPUEAR_ stubs, the branch in
1325 question is the one in the stub itself. */
1326 BFD_ASSERT (stub_type == nonovl_stub);
1327 g->br_addr = g->stub_addr;
1328 br_dest = to;
1331 bfd_put_32 (sec->owner, dest_ovl - 1,
1332 sec->contents + sec->size + 0);
1333 set_id = (dest_ovl - 1) >> htab->num_lines_log2;
1334 bfd_put_32 (sec->owner, (set_id << 18) | (dest & 0x3ffff),
1335 sec->contents + sec->size + 4);
1336 bfd_put_32 (sec->owner, (lrlive << 29) | (g->br_addr & 0x3ffff),
1337 sec->contents + sec->size + 8);
1338 bfd_put_32 (sec->owner, BRASL + ((to << 5) & 0x007fff80) + 75,
1339 sec->contents + sec->size + 12);
1340 patt = dest ^ br_dest;
1341 if (irela != NULL && ELF32_R_TYPE (irela->r_info) == R_SPU_REL16)
1342 patt = (dest - g->br_addr) ^ (br_dest - g->br_addr);
1343 bfd_put_32 (sec->owner, (patt << 5) & 0x007fff80,
1344 sec->contents + sec->size + 16 + (g->br_addr & 0xf));
1345 if (ovl == 0)
1346 /* Extra space for linked list entries. */
1347 sec->size += 16;
1348 break;
1350 default:
1351 abort ();
1353 sec->size += ovl_stub_size (htab->params->ovly_flavour);
1355 if (htab->params->emit_stub_syms)
1357 size_t len;
1358 char *name;
1359 int add;
1361 len = 8 + sizeof (".ovl_call.") - 1;
1362 if (h != NULL)
1363 len += strlen (h->root.root.string);
1364 else
1365 len += 8 + 1 + 8;
1366 add = 0;
1367 if (irela != NULL)
1368 add = (int) irela->r_addend & 0xffffffff;
1369 if (add != 0)
1370 len += 1 + 8;
1371 name = bfd_malloc (len);
1372 if (name == NULL)
1373 return FALSE;
1375 sprintf (name, "%08x.ovl_call.", g->ovl);
1376 if (h != NULL)
1377 strcpy (name + 8 + sizeof (".ovl_call.") - 1, h->root.root.string);
1378 else
1379 sprintf (name + 8 + sizeof (".ovl_call.") - 1, "%x:%x",
1380 dest_sec->id & 0xffffffff,
1381 (int) ELF32_R_SYM (irela->r_info) & 0xffffffff);
1382 if (add != 0)
1383 sprintf (name + len - 9, "+%x", add);
1385 h = elf_link_hash_lookup (&htab->elf, name, TRUE, TRUE, FALSE);
1386 free (name);
1387 if (h == NULL)
1388 return FALSE;
1389 if (h->root.type == bfd_link_hash_new)
1391 h->root.type = bfd_link_hash_defined;
1392 h->root.u.def.section = sec;
1393 h->size = ovl_stub_size (htab->params->ovly_flavour);
1394 h->root.u.def.value = sec->size - h->size;
1395 h->type = STT_FUNC;
1396 h->ref_regular = 1;
1397 h->def_regular = 1;
1398 h->ref_regular_nonweak = 1;
1399 h->forced_local = 1;
1400 h->non_elf = 0;
1404 return TRUE;
1407 /* Called via elf_link_hash_traverse to allocate stubs for any _SPUEAR_
1408 symbols. */
1410 static bfd_boolean
1411 allocate_spuear_stubs (struct elf_link_hash_entry *h, void *inf)
1413 /* Symbols starting with _SPUEAR_ need a stub because they may be
1414 invoked by the PPU. */
1415 struct bfd_link_info *info = inf;
1416 struct spu_link_hash_table *htab = spu_hash_table (info);
1417 asection *sym_sec;
1419 if ((h->root.type == bfd_link_hash_defined
1420 || h->root.type == bfd_link_hash_defweak)
1421 && h->def_regular
1422 && strncmp (h->root.root.string, "_SPUEAR_", 8) == 0
1423 && (sym_sec = h->root.u.def.section) != NULL
1424 && sym_sec->output_section != bfd_abs_section_ptr
1425 && spu_elf_section_data (sym_sec->output_section) != NULL
1426 && (spu_elf_section_data (sym_sec->output_section)->u.o.ovl_index != 0
1427 || htab->params->non_overlay_stubs))
1429 return count_stub (htab, NULL, NULL, nonovl_stub, h, NULL);
1432 return TRUE;
1435 static bfd_boolean
1436 build_spuear_stubs (struct elf_link_hash_entry *h, void *inf)
1438 /* Symbols starting with _SPUEAR_ need a stub because they may be
1439 invoked by the PPU. */
1440 struct bfd_link_info *info = inf;
1441 struct spu_link_hash_table *htab = spu_hash_table (info);
1442 asection *sym_sec;
1444 if ((h->root.type == bfd_link_hash_defined
1445 || h->root.type == bfd_link_hash_defweak)
1446 && h->def_regular
1447 && strncmp (h->root.root.string, "_SPUEAR_", 8) == 0
1448 && (sym_sec = h->root.u.def.section) != NULL
1449 && sym_sec->output_section != bfd_abs_section_ptr
1450 && spu_elf_section_data (sym_sec->output_section) != NULL
1451 && (spu_elf_section_data (sym_sec->output_section)->u.o.ovl_index != 0
1452 || htab->params->non_overlay_stubs))
1454 return build_stub (info, NULL, NULL, nonovl_stub, h, NULL,
1455 h->root.u.def.value, sym_sec);
1458 return TRUE;
1461 /* Size or build stubs. */
1463 static bfd_boolean
1464 process_stubs (struct bfd_link_info *info, bfd_boolean build)
1466 struct spu_link_hash_table *htab = spu_hash_table (info);
1467 bfd *ibfd;
1469 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
1471 extern const bfd_target bfd_elf32_spu_vec;
1472 Elf_Internal_Shdr *symtab_hdr;
1473 asection *isec;
1474 Elf_Internal_Sym *local_syms = NULL;
1476 if (ibfd->xvec != &bfd_elf32_spu_vec)
1477 continue;
1479 /* We'll need the symbol table in a second. */
1480 symtab_hdr = &elf_tdata (ibfd)->symtab_hdr;
1481 if (symtab_hdr->sh_info == 0)
1482 continue;
1484 /* Walk over each section attached to the input bfd. */
1485 for (isec = ibfd->sections; isec != NULL; isec = isec->next)
1487 Elf_Internal_Rela *internal_relocs, *irelaend, *irela;
1489 /* If there aren't any relocs, then there's nothing more to do. */
1490 if ((isec->flags & SEC_RELOC) == 0
1491 || isec->reloc_count == 0)
1492 continue;
1494 if (!maybe_needs_stubs (isec))
1495 continue;
1497 /* Get the relocs. */
1498 internal_relocs = _bfd_elf_link_read_relocs (ibfd, isec, NULL, NULL,
1499 info->keep_memory);
1500 if (internal_relocs == NULL)
1501 goto error_ret_free_local;
1503 /* Now examine each relocation. */
1504 irela = internal_relocs;
1505 irelaend = irela + isec->reloc_count;
1506 for (; irela < irelaend; irela++)
1508 enum elf_spu_reloc_type r_type;
1509 unsigned int r_indx;
1510 asection *sym_sec;
1511 Elf_Internal_Sym *sym;
1512 struct elf_link_hash_entry *h;
1513 enum _stub_type stub_type;
1515 r_type = ELF32_R_TYPE (irela->r_info);
1516 r_indx = ELF32_R_SYM (irela->r_info);
1518 if (r_type >= R_SPU_max)
1520 bfd_set_error (bfd_error_bad_value);
1521 error_ret_free_internal:
1522 if (elf_section_data (isec)->relocs != internal_relocs)
1523 free (internal_relocs);
1524 error_ret_free_local:
1525 if (local_syms != NULL
1526 && (symtab_hdr->contents
1527 != (unsigned char *) local_syms))
1528 free (local_syms);
1529 return FALSE;
1532 /* Determine the reloc target section. */
1533 if (!get_sym_h (&h, &sym, &sym_sec, &local_syms, r_indx, ibfd))
1534 goto error_ret_free_internal;
1536 stub_type = needs_ovl_stub (h, sym, sym_sec, isec, irela,
1537 NULL, info);
1538 if (stub_type == no_stub)
1539 continue;
1540 else if (stub_type == stub_error)
1541 goto error_ret_free_internal;
1543 if (htab->stub_count == NULL)
1545 bfd_size_type amt;
1546 amt = (htab->num_overlays + 1) * sizeof (*htab->stub_count);
1547 htab->stub_count = bfd_zmalloc (amt);
1548 if (htab->stub_count == NULL)
1549 goto error_ret_free_internal;
1552 if (!build)
1554 if (!count_stub (htab, ibfd, isec, stub_type, h, irela))
1555 goto error_ret_free_internal;
1557 else
1559 bfd_vma dest;
1561 if (h != NULL)
1562 dest = h->root.u.def.value;
1563 else
1564 dest = sym->st_value;
1565 dest += irela->r_addend;
1566 if (!build_stub (info, ibfd, isec, stub_type, h, irela,
1567 dest, sym_sec))
1568 goto error_ret_free_internal;
1572 /* We're done with the internal relocs, free them. */
1573 if (elf_section_data (isec)->relocs != internal_relocs)
1574 free (internal_relocs);
1577 if (local_syms != NULL
1578 && symtab_hdr->contents != (unsigned char *) local_syms)
1580 if (!info->keep_memory)
1581 free (local_syms);
1582 else
1583 symtab_hdr->contents = (unsigned char *) local_syms;
1587 return TRUE;
1590 /* Allocate space for overlay call and return stubs. */
1593 spu_elf_size_stubs (struct bfd_link_info *info)
1595 struct spu_link_hash_table *htab;
1596 bfd *ibfd;
1597 bfd_size_type amt;
1598 flagword flags;
1599 unsigned int i;
1600 asection *stub;
1601 const char *ovout;
1603 if (!process_stubs (info, FALSE))
1604 return 0;
1606 htab = spu_hash_table (info);
1607 elf_link_hash_traverse (&htab->elf, allocate_spuear_stubs, info);
1608 if (htab->stub_err)
1609 return 0;
1611 if (htab->stub_count == NULL)
1612 return 1;
1614 ibfd = info->input_bfds;
1615 amt = (htab->num_overlays + 1) * sizeof (*htab->stub_sec);
1616 htab->stub_sec = bfd_zmalloc (amt);
1617 if (htab->stub_sec == NULL)
1618 return 0;
1620 flags = (SEC_ALLOC | SEC_LOAD | SEC_CODE | SEC_READONLY
1621 | SEC_HAS_CONTENTS | SEC_IN_MEMORY);
1622 stub = bfd_make_section_anyway_with_flags (ibfd, ".stub", flags);
1623 htab->stub_sec[0] = stub;
1624 if (stub == NULL
1625 || !bfd_set_section_alignment (ibfd, stub,
1626 htab->params->ovly_flavour + 3))
1627 return 0;
1628 stub->size = htab->stub_count[0] * ovl_stub_size (htab->params->ovly_flavour);
1629 if (htab->params->ovly_flavour == ovly_soft_icache)
1630 /* Extra space for linked list entries. */
1631 stub->size += htab->stub_count[0] * 16;
1632 (*htab->params->place_spu_section) (stub, NULL, ".text");
1634 for (i = 0; i < htab->num_overlays; ++i)
1636 asection *osec = htab->ovl_sec[i];
1637 unsigned int ovl = spu_elf_section_data (osec)->u.o.ovl_index;
1638 stub = bfd_make_section_anyway_with_flags (ibfd, ".stub", flags);
1639 htab->stub_sec[ovl] = stub;
1640 if (stub == NULL
1641 || !bfd_set_section_alignment (ibfd, stub,
1642 htab->params->ovly_flavour + 3))
1643 return 0;
1644 stub->size = htab->stub_count[ovl] * ovl_stub_size (htab->params->ovly_flavour);
1645 (*htab->params->place_spu_section) (stub, osec, NULL);
1648 flags = (SEC_ALLOC | SEC_LOAD
1649 | SEC_HAS_CONTENTS | SEC_IN_MEMORY);
1650 htab->ovtab = bfd_make_section_anyway_with_flags (ibfd, ".ovtab", flags);
1651 if (htab->ovtab == NULL
1652 || !bfd_set_section_alignment (ibfd, htab->ovtab, 4))
1653 return 0;
1655 if (htab->params->ovly_flavour == ovly_soft_icache)
1657 /* Space for icache manager tables.
1658 a) Tag array, one quadword per cache line.
1659 b) Linked list elements, max_branch per line quadwords.
1660 c) Indirect branch descriptors, 8 quadwords. */
1661 htab->ovtab->size = 16 * (((1 + htab->params->max_branch)
1662 << htab->num_lines_log2)
1663 + 8);
1665 htab->init = bfd_make_section_anyway_with_flags (ibfd, ".ovini", flags);
1666 if (htab->init == NULL
1667 || !bfd_set_section_alignment (ibfd, htab->init, 4))
1668 return 0;
1670 htab->init->size = 16;
1671 (*htab->params->place_spu_section) (htab->init, NULL, ".ovl.init");
1673 else
1675 /* htab->ovtab consists of two arrays.
1676 . struct {
1677 . u32 vma;
1678 . u32 size;
1679 . u32 file_off;
1680 . u32 buf;
1681 . } _ovly_table[];
1683 . struct {
1684 . u32 mapped;
1685 . } _ovly_buf_table[];
1686 . */
1688 htab->ovtab->size = htab->num_overlays * 16 + 16 + htab->num_buf * 4;
1690 ovout = ".data";
1691 if (htab->params->ovly_flavour == ovly_soft_icache)
1692 ovout = ".data.icache";
1693 (*htab->params->place_spu_section) (htab->ovtab, NULL, ovout);
1695 htab->toe = bfd_make_section_anyway_with_flags (ibfd, ".toe", SEC_ALLOC);
1696 if (htab->toe == NULL
1697 || !bfd_set_section_alignment (ibfd, htab->toe, 4))
1698 return 0;
1699 htab->toe->size = htab->params->ovly_flavour == ovly_soft_icache ? 256 : 16;
1700 (*htab->params->place_spu_section) (htab->toe, NULL, ".toe");
1702 return 2;
1705 /* Functions to handle embedded spu_ovl.o object. */
1707 static void *
1708 ovl_mgr_open (struct bfd *nbfd ATTRIBUTE_UNUSED, void *stream)
1710 return stream;
1713 static file_ptr
1714 ovl_mgr_pread (struct bfd *abfd ATTRIBUTE_UNUSED,
1715 void *stream,
1716 void *buf,
1717 file_ptr nbytes,
1718 file_ptr offset)
1720 struct _ovl_stream *os;
1721 size_t count;
1722 size_t max;
1724 os = (struct _ovl_stream *) stream;
1725 max = (const char *) os->end - (const char *) os->start;
1727 if ((ufile_ptr) offset >= max)
1728 return 0;
1730 count = nbytes;
1731 if (count > max - offset)
1732 count = max - offset;
1734 memcpy (buf, (const char *) os->start + offset, count);
1735 return count;
1738 bfd_boolean
1739 spu_elf_open_builtin_lib (bfd **ovl_bfd, const struct _ovl_stream *stream)
1741 *ovl_bfd = bfd_openr_iovec ("builtin ovl_mgr",
1742 "elf32-spu",
1743 ovl_mgr_open,
1744 (void *) stream,
1745 ovl_mgr_pread,
1746 NULL,
1747 NULL);
1748 return *ovl_bfd != NULL;
1751 static unsigned int
1752 overlay_index (asection *sec)
1754 if (sec == NULL
1755 || sec->output_section == bfd_abs_section_ptr)
1756 return 0;
1757 return spu_elf_section_data (sec->output_section)->u.o.ovl_index;
1760 /* Define an STT_OBJECT symbol. */
1762 static struct elf_link_hash_entry *
1763 define_ovtab_symbol (struct spu_link_hash_table *htab, const char *name)
1765 struct elf_link_hash_entry *h;
1767 h = elf_link_hash_lookup (&htab->elf, name, TRUE, FALSE, FALSE);
1768 if (h == NULL)
1769 return NULL;
1771 if (h->root.type != bfd_link_hash_defined
1772 || !h->def_regular)
1774 h->root.type = bfd_link_hash_defined;
1775 h->root.u.def.section = htab->ovtab;
1776 h->type = STT_OBJECT;
1777 h->ref_regular = 1;
1778 h->def_regular = 1;
1779 h->ref_regular_nonweak = 1;
1780 h->non_elf = 0;
1782 else if (h->root.u.def.section->owner != NULL)
1784 (*_bfd_error_handler) (_("%B is not allowed to define %s"),
1785 h->root.u.def.section->owner,
1786 h->root.root.string);
1787 bfd_set_error (bfd_error_bad_value);
1788 return NULL;
1790 else
1792 (*_bfd_error_handler) (_("you are not allowed to define %s in a script"),
1793 h->root.root.string);
1794 bfd_set_error (bfd_error_bad_value);
1795 return NULL;
1798 return h;
1801 /* Fill in all stubs and the overlay tables. */
1803 static bfd_boolean
1804 spu_elf_build_stubs (struct bfd_link_info *info)
1806 struct spu_link_hash_table *htab = spu_hash_table (info);
1807 struct elf_link_hash_entry *h;
1808 bfd_byte *p;
1809 asection *s;
1810 bfd *obfd;
1811 unsigned int i;
1813 if (htab->stub_count == NULL)
1814 return TRUE;
1816 for (i = 0; i <= htab->num_overlays; i++)
1817 if (htab->stub_sec[i]->size != 0)
1819 htab->stub_sec[i]->contents = bfd_zalloc (htab->stub_sec[i]->owner,
1820 htab->stub_sec[i]->size);
1821 if (htab->stub_sec[i]->contents == NULL)
1822 return FALSE;
1823 htab->stub_sec[i]->rawsize = htab->stub_sec[i]->size;
1824 htab->stub_sec[i]->size = 0;
1827 h = htab->ovly_load;
1828 if (h == NULL)
1830 const char *ovly_mgr_entry = "__ovly_load";
1832 if (htab->params->ovly_flavour == ovly_soft_icache)
1833 ovly_mgr_entry = "__icache_br_handler";
1834 h = elf_link_hash_lookup (&htab->elf, ovly_mgr_entry,
1835 FALSE, FALSE, FALSE);
1836 htab->ovly_load = h;
1838 BFD_ASSERT (h != NULL
1839 && (h->root.type == bfd_link_hash_defined
1840 || h->root.type == bfd_link_hash_defweak)
1841 && h->def_regular);
1843 s = h->root.u.def.section->output_section;
1844 if (spu_elf_section_data (s)->u.o.ovl_index)
1846 (*_bfd_error_handler) (_("%s in overlay section"),
1847 h->root.root.string);
1848 bfd_set_error (bfd_error_bad_value);
1849 return FALSE;
1852 h = htab->ovly_return;
1853 if (h == NULL && htab->params->ovly_flavour != ovly_soft_icache)
1855 h = elf_link_hash_lookup (&htab->elf, "__ovly_return",
1856 FALSE, FALSE, FALSE);
1857 htab->ovly_return = h;
1860 /* Fill in all the stubs. */
1861 process_stubs (info, TRUE);
1862 if (!htab->stub_err)
1863 elf_link_hash_traverse (&htab->elf, build_spuear_stubs, info);
1865 if (htab->stub_err)
1867 (*_bfd_error_handler) (_("overlay stub relocation overflow"));
1868 bfd_set_error (bfd_error_bad_value);
1869 return FALSE;
1872 for (i = 0; i <= htab->num_overlays; i++)
1874 if (htab->stub_sec[i]->size != htab->stub_sec[i]->rawsize)
1876 (*_bfd_error_handler) (_("stubs don't match calculated size"));
1877 bfd_set_error (bfd_error_bad_value);
1878 return FALSE;
1880 htab->stub_sec[i]->rawsize = 0;
1883 if (htab->ovtab == NULL || htab->ovtab->size == 0)
1884 return TRUE;
1886 htab->ovtab->contents = bfd_zalloc (htab->ovtab->owner, htab->ovtab->size);
1887 if (htab->ovtab->contents == NULL)
1888 return FALSE;
1890 p = htab->ovtab->contents;
1891 if (htab->params->ovly_flavour == ovly_soft_icache)
1893 #define BI_HANDLER "__icache_ptr_handler0"
1894 char name[sizeof (BI_HANDLER)];
1895 bfd_vma off, icache_base, linklist, bihand;
1897 h = define_ovtab_symbol (htab, "__icache_tagbase");
1898 if (h == NULL)
1899 return FALSE;
1900 h->root.u.def.value = 0;
1901 h->size = 16 << htab->num_lines_log2;
1902 off = h->size;
1903 icache_base = htab->ovl_sec[0]->vma;
1904 linklist = (htab->ovtab->output_section->vma
1905 + htab->ovtab->output_offset
1906 + off);
1907 for (i = 0; i < htab->params->num_lines; i++)
1909 bfd_vma line_end = icache_base + ((i + 1) << htab->line_size_log2);
1910 bfd_vma stub_base = line_end - htab->params->max_branch * 32;
1911 bfd_vma link_elem = linklist + i * htab->params->max_branch * 16;
1912 bfd_vma locator = link_elem - stub_base / 2;
1914 bfd_put_32 (htab->ovtab->owner, locator, p + 4);
1915 bfd_put_16 (htab->ovtab->owner, link_elem, p + 8);
1916 bfd_put_16 (htab->ovtab->owner, link_elem, p + 10);
1917 bfd_put_16 (htab->ovtab->owner, link_elem, p + 12);
1918 bfd_put_16 (htab->ovtab->owner, link_elem, p + 14);
1919 p += 16;
1922 h = define_ovtab_symbol (htab, "__icache_linked_list");
1923 if (h == NULL)
1924 return FALSE;
1925 h->root.u.def.value = off;
1926 h->size = htab->params->max_branch << (htab->num_lines_log2 + 4);
1927 off += h->size;
1928 p += h->size;
1930 h = elf_link_hash_lookup (&htab->elf, "__icache_bi_handler",
1931 FALSE, FALSE, FALSE);
1932 bihand = 0;
1933 if (h != NULL
1934 && (h->root.type == bfd_link_hash_defined
1935 || h->root.type == bfd_link_hash_defweak)
1936 && h->def_regular)
1937 bihand = (h->root.u.def.value
1938 + h->root.u.def.section->output_offset
1939 + h->root.u.def.section->output_section->vma);
1940 memcpy (name, BI_HANDLER, sizeof (BI_HANDLER));
1941 for (i = 0; i < 8; i++)
1943 name[sizeof (BI_HANDLER) - 2] = '0' + i;
1944 h = define_ovtab_symbol (htab, name);
1945 if (h == NULL)
1946 return FALSE;
1947 h->root.u.def.value = off;
1948 h->size = 16;
1949 bfd_put_32 (htab->ovtab->owner, bihand, p);
1950 bfd_put_32 (htab->ovtab->owner, i << 28, p + 8);
1951 p += 16;
1952 off += 16;
1955 h = define_ovtab_symbol (htab, "__icache_base");
1956 if (h == NULL)
1957 return FALSE;
1958 h->root.u.def.value = htab->ovl_sec[0]->vma;
1959 h->root.u.def.section = bfd_abs_section_ptr;
1960 h->size = htab->num_buf << htab->line_size_log2;
1962 h = define_ovtab_symbol (htab, "__icache_neg_log2_linesize");
1963 if (h == NULL)
1964 return FALSE;
1965 h->root.u.def.value = -htab->line_size_log2;
1966 h->root.u.def.section = bfd_abs_section_ptr;
1968 if (htab->init != NULL && htab->init->size != 0)
1970 htab->init->contents = bfd_zalloc (htab->init->owner,
1971 htab->init->size);
1972 if (htab->init->contents == NULL)
1973 return FALSE;
1975 h = define_ovtab_symbol (htab, "__icache_fileoff");
1976 if (h == NULL)
1977 return FALSE;
1978 h->root.u.def.value = 0;
1979 h->root.u.def.section = htab->init;
1980 h->size = 8;
1983 else
1985 /* Write out _ovly_table. */
1986 /* set low bit of .size to mark non-overlay area as present. */
1987 p[7] = 1;
1988 obfd = htab->ovtab->output_section->owner;
1989 for (s = obfd->sections; s != NULL; s = s->next)
1991 unsigned int ovl_index = spu_elf_section_data (s)->u.o.ovl_index;
1993 if (ovl_index != 0)
1995 unsigned long off = ovl_index * 16;
1996 unsigned int ovl_buf = spu_elf_section_data (s)->u.o.ovl_buf;
1998 bfd_put_32 (htab->ovtab->owner, s->vma, p + off);
1999 bfd_put_32 (htab->ovtab->owner, (s->size + 15) & -16,
2000 p + off + 4);
2001 /* file_off written later in spu_elf_modify_program_headers. */
2002 bfd_put_32 (htab->ovtab->owner, ovl_buf, p + off + 12);
2006 h = define_ovtab_symbol (htab, "_ovly_table");
2007 if (h == NULL)
2008 return FALSE;
2009 h->root.u.def.value = 16;
2010 h->size = htab->num_overlays * 16;
2012 h = define_ovtab_symbol (htab, "_ovly_table_end");
2013 if (h == NULL)
2014 return FALSE;
2015 h->root.u.def.value = htab->num_overlays * 16 + 16;
2016 h->size = 0;
2018 h = define_ovtab_symbol (htab, "_ovly_buf_table");
2019 if (h == NULL)
2020 return FALSE;
2021 h->root.u.def.value = htab->num_overlays * 16 + 16;
2022 h->size = htab->num_buf * 4;
2024 h = define_ovtab_symbol (htab, "_ovly_buf_table_end");
2025 if (h == NULL)
2026 return FALSE;
2027 h->root.u.def.value = htab->num_overlays * 16 + 16 + htab->num_buf * 4;
2028 h->size = 0;
2031 h = define_ovtab_symbol (htab, "_EAR_");
2032 if (h == NULL)
2033 return FALSE;
2034 h->root.u.def.section = htab->toe;
2035 h->root.u.def.value = 0;
2036 h->size = htab->params->ovly_flavour == ovly_soft_icache ? 16 * 16 : 16;
2038 return TRUE;
2041 /* Check that all loadable section VMAs lie in the range
2042 LO .. HI inclusive, and stash some parameters for --auto-overlay. */
2044 asection *
2045 spu_elf_check_vma (struct bfd_link_info *info)
2047 struct elf_segment_map *m;
2048 unsigned int i;
2049 struct spu_link_hash_table *htab = spu_hash_table (info);
2050 bfd *abfd = info->output_bfd;
2051 bfd_vma hi = htab->params->local_store_hi;
2052 bfd_vma lo = htab->params->local_store_lo;
2054 htab->local_store = hi + 1 - lo;
2056 for (m = elf_tdata (abfd)->segment_map; m != NULL; m = m->next)
2057 if (m->p_type == PT_LOAD)
2058 for (i = 0; i < m->count; i++)
2059 if (m->sections[i]->size != 0
2060 && (m->sections[i]->vma < lo
2061 || m->sections[i]->vma > hi
2062 || m->sections[i]->vma + m->sections[i]->size - 1 > hi))
2063 return m->sections[i];
2065 /* No need for overlays if it all fits. */
2066 if (htab->params->ovly_flavour != ovly_soft_icache)
2067 htab->params->auto_overlay = 0;
2068 return NULL;
2071 /* OFFSET in SEC (presumably) is the beginning of a function prologue.
2072 Search for stack adjusting insns, and return the sp delta.
2073 If a store of lr is found save the instruction offset to *LR_STORE.
2074 If a stack adjusting instruction is found, save that offset to
2075 *SP_ADJUST. */
2077 static int
2078 find_function_stack_adjust (asection *sec,
2079 bfd_vma offset,
2080 bfd_vma *lr_store,
2081 bfd_vma *sp_adjust)
2083 int reg[128];
2085 memset (reg, 0, sizeof (reg));
2086 for ( ; offset + 4 <= sec->size; offset += 4)
2088 unsigned char buf[4];
2089 int rt, ra;
2090 int imm;
2092 /* Assume no relocs on stack adjusing insns. */
2093 if (!bfd_get_section_contents (sec->owner, sec, buf, offset, 4))
2094 break;
2096 rt = buf[3] & 0x7f;
2097 ra = ((buf[2] & 0x3f) << 1) | (buf[3] >> 7);
2099 if (buf[0] == 0x24 /* stqd */)
2101 if (rt == 0 /* lr */ && ra == 1 /* sp */)
2102 *lr_store = offset;
2103 continue;
2106 /* Partly decoded immediate field. */
2107 imm = (buf[1] << 9) | (buf[2] << 1) | (buf[3] >> 7);
2109 if (buf[0] == 0x1c /* ai */)
2111 imm >>= 7;
2112 imm = (imm ^ 0x200) - 0x200;
2113 reg[rt] = reg[ra] + imm;
2115 if (rt == 1 /* sp */)
2117 if (reg[rt] > 0)
2118 break;
2119 *sp_adjust = offset;
2120 return reg[rt];
2123 else if (buf[0] == 0x18 && (buf[1] & 0xe0) == 0 /* a */)
2125 int rb = ((buf[1] & 0x1f) << 2) | ((buf[2] & 0xc0) >> 6);
2127 reg[rt] = reg[ra] + reg[rb];
2128 if (rt == 1)
2130 if (reg[rt] > 0)
2131 break;
2132 *sp_adjust = offset;
2133 return reg[rt];
2136 else if (buf[0] == 0x08 && (buf[1] & 0xe0) == 0 /* sf */)
2138 int rb = ((buf[1] & 0x1f) << 2) | ((buf[2] & 0xc0) >> 6);
2140 reg[rt] = reg[rb] - reg[ra];
2141 if (rt == 1)
2143 if (reg[rt] > 0)
2144 break;
2145 *sp_adjust = offset;
2146 return reg[rt];
2149 else if ((buf[0] & 0xfc) == 0x40 /* il, ilh, ilhu, ila */)
2151 if (buf[0] >= 0x42 /* ila */)
2152 imm |= (buf[0] & 1) << 17;
2153 else
2155 imm &= 0xffff;
2157 if (buf[0] == 0x40 /* il */)
2159 if ((buf[1] & 0x80) == 0)
2160 continue;
2161 imm = (imm ^ 0x8000) - 0x8000;
2163 else if ((buf[1] & 0x80) == 0 /* ilhu */)
2164 imm <<= 16;
2166 reg[rt] = imm;
2167 continue;
2169 else if (buf[0] == 0x60 && (buf[1] & 0x80) != 0 /* iohl */)
2171 reg[rt] |= imm & 0xffff;
2172 continue;
2174 else if (buf[0] == 0x04 /* ori */)
2176 imm >>= 7;
2177 imm = (imm ^ 0x200) - 0x200;
2178 reg[rt] = reg[ra] | imm;
2179 continue;
2181 else if (buf[0] == 0x32 && (buf[1] & 0x80) != 0 /* fsmbi */)
2183 reg[rt] = ( ((imm & 0x8000) ? 0xff000000 : 0)
2184 | ((imm & 0x4000) ? 0x00ff0000 : 0)
2185 | ((imm & 0x2000) ? 0x0000ff00 : 0)
2186 | ((imm & 0x1000) ? 0x000000ff : 0));
2187 continue;
2189 else if (buf[0] == 0x16 /* andbi */)
2191 imm >>= 7;
2192 imm &= 0xff;
2193 imm |= imm << 8;
2194 imm |= imm << 16;
2195 reg[rt] = reg[ra] & imm;
2196 continue;
2198 else if (buf[0] == 0x33 && imm == 1 /* brsl .+4 */)
2200 /* Used in pic reg load. Say rt is trashed. Won't be used
2201 in stack adjust, but we need to continue past this branch. */
2202 reg[rt] = 0;
2203 continue;
2205 else if (is_branch (buf) || is_indirect_branch (buf))
2206 /* If we hit a branch then we must be out of the prologue. */
2207 break;
2210 return 0;
2213 /* qsort predicate to sort symbols by section and value. */
2215 static Elf_Internal_Sym *sort_syms_syms;
2216 static asection **sort_syms_psecs;
2218 static int
2219 sort_syms (const void *a, const void *b)
2221 Elf_Internal_Sym *const *s1 = a;
2222 Elf_Internal_Sym *const *s2 = b;
2223 asection *sec1,*sec2;
2224 bfd_signed_vma delta;
2226 sec1 = sort_syms_psecs[*s1 - sort_syms_syms];
2227 sec2 = sort_syms_psecs[*s2 - sort_syms_syms];
2229 if (sec1 != sec2)
2230 return sec1->index - sec2->index;
2232 delta = (*s1)->st_value - (*s2)->st_value;
2233 if (delta != 0)
2234 return delta < 0 ? -1 : 1;
2236 delta = (*s2)->st_size - (*s1)->st_size;
2237 if (delta != 0)
2238 return delta < 0 ? -1 : 1;
2240 return *s1 < *s2 ? -1 : 1;
2243 /* Allocate a struct spu_elf_stack_info with MAX_FUN struct function_info
2244 entries for section SEC. */
2246 static struct spu_elf_stack_info *
2247 alloc_stack_info (asection *sec, int max_fun)
2249 struct _spu_elf_section_data *sec_data = spu_elf_section_data (sec);
2250 bfd_size_type amt;
2252 amt = sizeof (struct spu_elf_stack_info);
2253 amt += (max_fun - 1) * sizeof (struct function_info);
2254 sec_data->u.i.stack_info = bfd_zmalloc (amt);
2255 if (sec_data->u.i.stack_info != NULL)
2256 sec_data->u.i.stack_info->max_fun = max_fun;
2257 return sec_data->u.i.stack_info;
2260 /* Add a new struct function_info describing a (part of a) function
2261 starting at SYM_H. Keep the array sorted by address. */
2263 static struct function_info *
2264 maybe_insert_function (asection *sec,
2265 void *sym_h,
2266 bfd_boolean global,
2267 bfd_boolean is_func)
2269 struct _spu_elf_section_data *sec_data = spu_elf_section_data (sec);
2270 struct spu_elf_stack_info *sinfo = sec_data->u.i.stack_info;
2271 int i;
2272 bfd_vma off, size;
2274 if (sinfo == NULL)
2276 sinfo = alloc_stack_info (sec, 20);
2277 if (sinfo == NULL)
2278 return NULL;
2281 if (!global)
2283 Elf_Internal_Sym *sym = sym_h;
2284 off = sym->st_value;
2285 size = sym->st_size;
2287 else
2289 struct elf_link_hash_entry *h = sym_h;
2290 off = h->root.u.def.value;
2291 size = h->size;
2294 for (i = sinfo->num_fun; --i >= 0; )
2295 if (sinfo->fun[i].lo <= off)
2296 break;
2298 if (i >= 0)
2300 /* Don't add another entry for an alias, but do update some
2301 info. */
2302 if (sinfo->fun[i].lo == off)
2304 /* Prefer globals over local syms. */
2305 if (global && !sinfo->fun[i].global)
2307 sinfo->fun[i].global = TRUE;
2308 sinfo->fun[i].u.h = sym_h;
2310 if (is_func)
2311 sinfo->fun[i].is_func = TRUE;
2312 return &sinfo->fun[i];
2314 /* Ignore a zero-size symbol inside an existing function. */
2315 else if (sinfo->fun[i].hi > off && size == 0)
2316 return &sinfo->fun[i];
2319 if (sinfo->num_fun >= sinfo->max_fun)
2321 bfd_size_type amt = sizeof (struct spu_elf_stack_info);
2322 bfd_size_type old = amt;
2324 old += (sinfo->max_fun - 1) * sizeof (struct function_info);
2325 sinfo->max_fun += 20 + (sinfo->max_fun >> 1);
2326 amt += (sinfo->max_fun - 1) * sizeof (struct function_info);
2327 sinfo = bfd_realloc (sinfo, amt);
2328 if (sinfo == NULL)
2329 return NULL;
2330 memset ((char *) sinfo + old, 0, amt - old);
2331 sec_data->u.i.stack_info = sinfo;
2334 if (++i < sinfo->num_fun)
2335 memmove (&sinfo->fun[i + 1], &sinfo->fun[i],
2336 (sinfo->num_fun - i) * sizeof (sinfo->fun[i]));
2337 sinfo->fun[i].is_func = is_func;
2338 sinfo->fun[i].global = global;
2339 sinfo->fun[i].sec = sec;
2340 if (global)
2341 sinfo->fun[i].u.h = sym_h;
2342 else
2343 sinfo->fun[i].u.sym = sym_h;
2344 sinfo->fun[i].lo = off;
2345 sinfo->fun[i].hi = off + size;
2346 sinfo->fun[i].lr_store = -1;
2347 sinfo->fun[i].sp_adjust = -1;
2348 sinfo->fun[i].stack = -find_function_stack_adjust (sec, off,
2349 &sinfo->fun[i].lr_store,
2350 &sinfo->fun[i].sp_adjust);
2351 sinfo->num_fun += 1;
2352 return &sinfo->fun[i];
2355 /* Return the name of FUN. */
2357 static const char *
2358 func_name (struct function_info *fun)
2360 asection *sec;
2361 bfd *ibfd;
2362 Elf_Internal_Shdr *symtab_hdr;
2364 while (fun->start != NULL)
2365 fun = fun->start;
2367 if (fun->global)
2368 return fun->u.h->root.root.string;
2370 sec = fun->sec;
2371 if (fun->u.sym->st_name == 0)
2373 size_t len = strlen (sec->name);
2374 char *name = bfd_malloc (len + 10);
2375 if (name == NULL)
2376 return "(null)";
2377 sprintf (name, "%s+%lx", sec->name,
2378 (unsigned long) fun->u.sym->st_value & 0xffffffff);
2379 return name;
2381 ibfd = sec->owner;
2382 symtab_hdr = &elf_tdata (ibfd)->symtab_hdr;
2383 return bfd_elf_sym_name (ibfd, symtab_hdr, fun->u.sym, sec);
2386 /* Read the instruction at OFF in SEC. Return true iff the instruction
2387 is a nop, lnop, or stop 0 (all zero insn). */
2389 static bfd_boolean
2390 is_nop (asection *sec, bfd_vma off)
2392 unsigned char insn[4];
2394 if (off + 4 > sec->size
2395 || !bfd_get_section_contents (sec->owner, sec, insn, off, 4))
2396 return FALSE;
2397 if ((insn[0] & 0xbf) == 0 && (insn[1] & 0xe0) == 0x20)
2398 return TRUE;
2399 if (insn[0] == 0 && insn[1] == 0 && insn[2] == 0 && insn[3] == 0)
2400 return TRUE;
2401 return FALSE;
2404 /* Extend the range of FUN to cover nop padding up to LIMIT.
2405 Return TRUE iff some instruction other than a NOP was found. */
2407 static bfd_boolean
2408 insns_at_end (struct function_info *fun, bfd_vma limit)
2410 bfd_vma off = (fun->hi + 3) & -4;
2412 while (off < limit && is_nop (fun->sec, off))
2413 off += 4;
2414 if (off < limit)
2416 fun->hi = off;
2417 return TRUE;
2419 fun->hi = limit;
2420 return FALSE;
2423 /* Check and fix overlapping function ranges. Return TRUE iff there
2424 are gaps in the current info we have about functions in SEC. */
2426 static bfd_boolean
2427 check_function_ranges (asection *sec, struct bfd_link_info *info)
2429 struct _spu_elf_section_data *sec_data = spu_elf_section_data (sec);
2430 struct spu_elf_stack_info *sinfo = sec_data->u.i.stack_info;
2431 int i;
2432 bfd_boolean gaps = FALSE;
2434 if (sinfo == NULL)
2435 return FALSE;
2437 for (i = 1; i < sinfo->num_fun; i++)
2438 if (sinfo->fun[i - 1].hi > sinfo->fun[i].lo)
2440 /* Fix overlapping symbols. */
2441 const char *f1 = func_name (&sinfo->fun[i - 1]);
2442 const char *f2 = func_name (&sinfo->fun[i]);
2444 info->callbacks->einfo (_("warning: %s overlaps %s\n"), f1, f2);
2445 sinfo->fun[i - 1].hi = sinfo->fun[i].lo;
2447 else if (insns_at_end (&sinfo->fun[i - 1], sinfo->fun[i].lo))
2448 gaps = TRUE;
2450 if (sinfo->num_fun == 0)
2451 gaps = TRUE;
2452 else
2454 if (sinfo->fun[0].lo != 0)
2455 gaps = TRUE;
2456 if (sinfo->fun[sinfo->num_fun - 1].hi > sec->size)
2458 const char *f1 = func_name (&sinfo->fun[sinfo->num_fun - 1]);
2460 info->callbacks->einfo (_("warning: %s exceeds section size\n"), f1);
2461 sinfo->fun[sinfo->num_fun - 1].hi = sec->size;
2463 else if (insns_at_end (&sinfo->fun[sinfo->num_fun - 1], sec->size))
2464 gaps = TRUE;
2466 return gaps;
2469 /* Search current function info for a function that contains address
2470 OFFSET in section SEC. */
2472 static struct function_info *
2473 find_function (asection *sec, bfd_vma offset, struct bfd_link_info *info)
2475 struct _spu_elf_section_data *sec_data = spu_elf_section_data (sec);
2476 struct spu_elf_stack_info *sinfo = sec_data->u.i.stack_info;
2477 int lo, hi, mid;
2479 lo = 0;
2480 hi = sinfo->num_fun;
2481 while (lo < hi)
2483 mid = (lo + hi) / 2;
2484 if (offset < sinfo->fun[mid].lo)
2485 hi = mid;
2486 else if (offset >= sinfo->fun[mid].hi)
2487 lo = mid + 1;
2488 else
2489 return &sinfo->fun[mid];
2491 info->callbacks->einfo (_("%A:0x%v not found in function table\n"),
2492 sec, offset);
2493 bfd_set_error (bfd_error_bad_value);
2494 return NULL;
2497 /* Add CALLEE to CALLER call list if not already present. Return TRUE
2498 if CALLEE was new. If this function return FALSE, CALLEE should
2499 be freed. */
2501 static bfd_boolean
2502 insert_callee (struct function_info *caller, struct call_info *callee)
2504 struct call_info **pp, *p;
2506 for (pp = &caller->call_list; (p = *pp) != NULL; pp = &p->next)
2507 if (p->fun == callee->fun)
2509 /* Tail calls use less stack than normal calls. Retain entry
2510 for normal call over one for tail call. */
2511 p->is_tail &= callee->is_tail;
2512 if (!p->is_tail)
2514 p->fun->start = NULL;
2515 p->fun->is_func = TRUE;
2517 p->count += 1;
2518 /* Reorder list so most recent call is first. */
2519 *pp = p->next;
2520 p->next = caller->call_list;
2521 caller->call_list = p;
2522 return FALSE;
2524 callee->next = caller->call_list;
2525 callee->count += 1;
2526 caller->call_list = callee;
2527 return TRUE;
2530 /* Copy CALL and insert the copy into CALLER. */
2532 static bfd_boolean
2533 copy_callee (struct function_info *caller, const struct call_info *call)
2535 struct call_info *callee;
2536 callee = bfd_malloc (sizeof (*callee));
2537 if (callee == NULL)
2538 return FALSE;
2539 *callee = *call;
2540 if (!insert_callee (caller, callee))
2541 free (callee);
2542 return TRUE;
2545 /* We're only interested in code sections. Testing SEC_IN_MEMORY excludes
2546 overlay stub sections. */
2548 static bfd_boolean
2549 interesting_section (asection *s)
2551 return (s->output_section != bfd_abs_section_ptr
2552 && ((s->flags & (SEC_ALLOC | SEC_LOAD | SEC_CODE | SEC_IN_MEMORY))
2553 == (SEC_ALLOC | SEC_LOAD | SEC_CODE))
2554 && s->size != 0);
2557 /* Rummage through the relocs for SEC, looking for function calls.
2558 If CALL_TREE is true, fill in call graph. If CALL_TREE is false,
2559 mark destination symbols on calls as being functions. Also
2560 look at branches, which may be tail calls or go to hot/cold
2561 section part of same function. */
2563 static bfd_boolean
2564 mark_functions_via_relocs (asection *sec,
2565 struct bfd_link_info *info,
2566 int call_tree)
2568 Elf_Internal_Rela *internal_relocs, *irelaend, *irela;
2569 Elf_Internal_Shdr *symtab_hdr;
2570 void *psyms;
2571 unsigned int priority = 0;
2572 static bfd_boolean warned;
2574 if (!interesting_section (sec)
2575 || sec->reloc_count == 0)
2576 return TRUE;
2578 internal_relocs = _bfd_elf_link_read_relocs (sec->owner, sec, NULL, NULL,
2579 info->keep_memory);
2580 if (internal_relocs == NULL)
2581 return FALSE;
2583 symtab_hdr = &elf_tdata (sec->owner)->symtab_hdr;
2584 psyms = &symtab_hdr->contents;
2585 irela = internal_relocs;
2586 irelaend = irela + sec->reloc_count;
2587 for (; irela < irelaend; irela++)
2589 enum elf_spu_reloc_type r_type;
2590 unsigned int r_indx;
2591 asection *sym_sec;
2592 Elf_Internal_Sym *sym;
2593 struct elf_link_hash_entry *h;
2594 bfd_vma val;
2595 bfd_boolean reject, is_call;
2596 struct function_info *caller;
2597 struct call_info *callee;
2599 reject = FALSE;
2600 r_type = ELF32_R_TYPE (irela->r_info);
2601 if (r_type != R_SPU_REL16
2602 && r_type != R_SPU_ADDR16)
2604 reject = TRUE;
2605 if (!(call_tree && spu_hash_table (info)->params->auto_overlay))
2606 continue;
2609 r_indx = ELF32_R_SYM (irela->r_info);
2610 if (!get_sym_h (&h, &sym, &sym_sec, psyms, r_indx, sec->owner))
2611 return FALSE;
2613 if (sym_sec == NULL
2614 || sym_sec->output_section == bfd_abs_section_ptr)
2615 continue;
2617 is_call = FALSE;
2618 if (!reject)
2620 unsigned char insn[4];
2622 if (!bfd_get_section_contents (sec->owner, sec, insn,
2623 irela->r_offset, 4))
2624 return FALSE;
2625 if (is_branch (insn))
2627 is_call = (insn[0] & 0xfd) == 0x31;
2628 priority = insn[1] & 0x0f;
2629 priority <<= 8;
2630 priority |= insn[2];
2631 priority <<= 8;
2632 priority |= insn[3];
2633 priority >>= 7;
2634 if ((sym_sec->flags & (SEC_ALLOC | SEC_LOAD | SEC_CODE))
2635 != (SEC_ALLOC | SEC_LOAD | SEC_CODE))
2637 if (!warned)
2638 info->callbacks->einfo
2639 (_("%B(%A+0x%v): call to non-code section"
2640 " %B(%A), analysis incomplete\n"),
2641 sec->owner, sec, irela->r_offset,
2642 sym_sec->owner, sym_sec);
2643 warned = TRUE;
2644 continue;
2647 else
2649 reject = TRUE;
2650 if (!(call_tree && spu_hash_table (info)->params->auto_overlay)
2651 || is_hint (insn))
2652 continue;
2656 if (reject)
2658 /* For --auto-overlay, count possible stubs we need for
2659 function pointer references. */
2660 unsigned int sym_type;
2661 if (h)
2662 sym_type = h->type;
2663 else
2664 sym_type = ELF_ST_TYPE (sym->st_info);
2665 if (sym_type == STT_FUNC)
2666 spu_hash_table (info)->non_ovly_stub += 1;
2667 continue;
2670 if (h)
2671 val = h->root.u.def.value;
2672 else
2673 val = sym->st_value;
2674 val += irela->r_addend;
2676 if (!call_tree)
2678 struct function_info *fun;
2680 if (irela->r_addend != 0)
2682 Elf_Internal_Sym *fake = bfd_zmalloc (sizeof (*fake));
2683 if (fake == NULL)
2684 return FALSE;
2685 fake->st_value = val;
2686 fake->st_shndx
2687 = _bfd_elf_section_from_bfd_section (sym_sec->owner, sym_sec);
2688 sym = fake;
2690 if (sym)
2691 fun = maybe_insert_function (sym_sec, sym, FALSE, is_call);
2692 else
2693 fun = maybe_insert_function (sym_sec, h, TRUE, is_call);
2694 if (fun == NULL)
2695 return FALSE;
2696 if (irela->r_addend != 0
2697 && fun->u.sym != sym)
2698 free (sym);
2699 continue;
2702 caller = find_function (sec, irela->r_offset, info);
2703 if (caller == NULL)
2704 return FALSE;
2705 callee = bfd_malloc (sizeof *callee);
2706 if (callee == NULL)
2707 return FALSE;
2709 callee->fun = find_function (sym_sec, val, info);
2710 if (callee->fun == NULL)
2711 return FALSE;
2712 callee->is_tail = !is_call;
2713 callee->is_pasted = FALSE;
2714 callee->priority = priority;
2715 callee->count = 0;
2716 if (callee->fun->last_caller != sec)
2718 callee->fun->last_caller = sec;
2719 callee->fun->call_count += 1;
2721 if (!insert_callee (caller, callee))
2722 free (callee);
2723 else if (!is_call
2724 && !callee->fun->is_func
2725 && callee->fun->stack == 0)
2727 /* This is either a tail call or a branch from one part of
2728 the function to another, ie. hot/cold section. If the
2729 destination has been called by some other function then
2730 it is a separate function. We also assume that functions
2731 are not split across input files. */
2732 if (sec->owner != sym_sec->owner)
2734 callee->fun->start = NULL;
2735 callee->fun->is_func = TRUE;
2737 else if (callee->fun->start == NULL)
2738 callee->fun->start = caller;
2739 else
2741 struct function_info *callee_start;
2742 struct function_info *caller_start;
2743 callee_start = callee->fun;
2744 while (callee_start->start)
2745 callee_start = callee_start->start;
2746 caller_start = caller;
2747 while (caller_start->start)
2748 caller_start = caller_start->start;
2749 if (caller_start != callee_start)
2751 callee->fun->start = NULL;
2752 callee->fun->is_func = TRUE;
2758 return TRUE;
2761 /* Handle something like .init or .fini, which has a piece of a function.
2762 These sections are pasted together to form a single function. */
2764 static bfd_boolean
2765 pasted_function (asection *sec)
2767 struct bfd_link_order *l;
2768 struct _spu_elf_section_data *sec_data;
2769 struct spu_elf_stack_info *sinfo;
2770 Elf_Internal_Sym *fake;
2771 struct function_info *fun, *fun_start;
2773 fake = bfd_zmalloc (sizeof (*fake));
2774 if (fake == NULL)
2775 return FALSE;
2776 fake->st_value = 0;
2777 fake->st_size = sec->size;
2778 fake->st_shndx
2779 = _bfd_elf_section_from_bfd_section (sec->owner, sec);
2780 fun = maybe_insert_function (sec, fake, FALSE, FALSE);
2781 if (!fun)
2782 return FALSE;
2784 /* Find a function immediately preceding this section. */
2785 fun_start = NULL;
2786 for (l = sec->output_section->map_head.link_order; l != NULL; l = l->next)
2788 if (l->u.indirect.section == sec)
2790 if (fun_start != NULL)
2792 struct call_info *callee = bfd_malloc (sizeof *callee);
2793 if (callee == NULL)
2794 return FALSE;
2796 fun->start = fun_start;
2797 callee->fun = fun;
2798 callee->is_tail = TRUE;
2799 callee->is_pasted = TRUE;
2800 callee->count = 0;
2801 if (!insert_callee (fun_start, callee))
2802 free (callee);
2803 return TRUE;
2805 break;
2807 if (l->type == bfd_indirect_link_order
2808 && (sec_data = spu_elf_section_data (l->u.indirect.section)) != NULL
2809 && (sinfo = sec_data->u.i.stack_info) != NULL
2810 && sinfo->num_fun != 0)
2811 fun_start = &sinfo->fun[sinfo->num_fun - 1];
2814 /* Don't return an error if we did not find a function preceding this
2815 section. The section may have incorrect flags. */
2816 return TRUE;
2819 /* Map address ranges in code sections to functions. */
2821 static bfd_boolean
2822 discover_functions (struct bfd_link_info *info)
2824 bfd *ibfd;
2825 int bfd_idx;
2826 Elf_Internal_Sym ***psym_arr;
2827 asection ***sec_arr;
2828 bfd_boolean gaps = FALSE;
2830 bfd_idx = 0;
2831 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
2832 bfd_idx++;
2834 psym_arr = bfd_zmalloc (bfd_idx * sizeof (*psym_arr));
2835 if (psym_arr == NULL)
2836 return FALSE;
2837 sec_arr = bfd_zmalloc (bfd_idx * sizeof (*sec_arr));
2838 if (sec_arr == NULL)
2839 return FALSE;
2841 for (ibfd = info->input_bfds, bfd_idx = 0;
2842 ibfd != NULL;
2843 ibfd = ibfd->link_next, bfd_idx++)
2845 extern const bfd_target bfd_elf32_spu_vec;
2846 Elf_Internal_Shdr *symtab_hdr;
2847 asection *sec;
2848 size_t symcount;
2849 Elf_Internal_Sym *syms, *sy, **psyms, **psy;
2850 asection **psecs, **p;
2852 if (ibfd->xvec != &bfd_elf32_spu_vec)
2853 continue;
2855 /* Read all the symbols. */
2856 symtab_hdr = &elf_tdata (ibfd)->symtab_hdr;
2857 symcount = symtab_hdr->sh_size / symtab_hdr->sh_entsize;
2858 if (symcount == 0)
2860 if (!gaps)
2861 for (sec = ibfd->sections; sec != NULL && !gaps; sec = sec->next)
2862 if (interesting_section (sec))
2864 gaps = TRUE;
2865 break;
2867 continue;
2870 if (symtab_hdr->contents != NULL)
2872 /* Don't use cached symbols since the generic ELF linker
2873 code only reads local symbols, and we need globals too. */
2874 free (symtab_hdr->contents);
2875 symtab_hdr->contents = NULL;
2877 syms = bfd_elf_get_elf_syms (ibfd, symtab_hdr, symcount, 0,
2878 NULL, NULL, NULL);
2879 symtab_hdr->contents = (void *) syms;
2880 if (syms == NULL)
2881 return FALSE;
2883 /* Select defined function symbols that are going to be output. */
2884 psyms = bfd_malloc ((symcount + 1) * sizeof (*psyms));
2885 if (psyms == NULL)
2886 return FALSE;
2887 psym_arr[bfd_idx] = psyms;
2888 psecs = bfd_malloc (symcount * sizeof (*psecs));
2889 if (psecs == NULL)
2890 return FALSE;
2891 sec_arr[bfd_idx] = psecs;
2892 for (psy = psyms, p = psecs, sy = syms; sy < syms + symcount; ++p, ++sy)
2893 if (ELF_ST_TYPE (sy->st_info) == STT_NOTYPE
2894 || ELF_ST_TYPE (sy->st_info) == STT_FUNC)
2896 asection *s;
2898 *p = s = bfd_section_from_elf_index (ibfd, sy->st_shndx);
2899 if (s != NULL && interesting_section (s))
2900 *psy++ = sy;
2902 symcount = psy - psyms;
2903 *psy = NULL;
2905 /* Sort them by section and offset within section. */
2906 sort_syms_syms = syms;
2907 sort_syms_psecs = psecs;
2908 qsort (psyms, symcount, sizeof (*psyms), sort_syms);
2910 /* Now inspect the function symbols. */
2911 for (psy = psyms; psy < psyms + symcount; )
2913 asection *s = psecs[*psy - syms];
2914 Elf_Internal_Sym **psy2;
2916 for (psy2 = psy; ++psy2 < psyms + symcount; )
2917 if (psecs[*psy2 - syms] != s)
2918 break;
2920 if (!alloc_stack_info (s, psy2 - psy))
2921 return FALSE;
2922 psy = psy2;
2925 /* First install info about properly typed and sized functions.
2926 In an ideal world this will cover all code sections, except
2927 when partitioning functions into hot and cold sections,
2928 and the horrible pasted together .init and .fini functions. */
2929 for (psy = psyms; psy < psyms + symcount; ++psy)
2931 sy = *psy;
2932 if (ELF_ST_TYPE (sy->st_info) == STT_FUNC)
2934 asection *s = psecs[sy - syms];
2935 if (!maybe_insert_function (s, sy, FALSE, TRUE))
2936 return FALSE;
2940 for (sec = ibfd->sections; sec != NULL && !gaps; sec = sec->next)
2941 if (interesting_section (sec))
2942 gaps |= check_function_ranges (sec, info);
2945 if (gaps)
2947 /* See if we can discover more function symbols by looking at
2948 relocations. */
2949 for (ibfd = info->input_bfds, bfd_idx = 0;
2950 ibfd != NULL;
2951 ibfd = ibfd->link_next, bfd_idx++)
2953 asection *sec;
2955 if (psym_arr[bfd_idx] == NULL)
2956 continue;
2958 for (sec = ibfd->sections; sec != NULL; sec = sec->next)
2959 if (!mark_functions_via_relocs (sec, info, FALSE))
2960 return FALSE;
2963 for (ibfd = info->input_bfds, bfd_idx = 0;
2964 ibfd != NULL;
2965 ibfd = ibfd->link_next, bfd_idx++)
2967 Elf_Internal_Shdr *symtab_hdr;
2968 asection *sec;
2969 Elf_Internal_Sym *syms, *sy, **psyms, **psy;
2970 asection **psecs;
2972 if ((psyms = psym_arr[bfd_idx]) == NULL)
2973 continue;
2975 psecs = sec_arr[bfd_idx];
2977 symtab_hdr = &elf_tdata (ibfd)->symtab_hdr;
2978 syms = (Elf_Internal_Sym *) symtab_hdr->contents;
2980 gaps = FALSE;
2981 for (sec = ibfd->sections; sec != NULL && !gaps; sec = sec->next)
2982 if (interesting_section (sec))
2983 gaps |= check_function_ranges (sec, info);
2984 if (!gaps)
2985 continue;
2987 /* Finally, install all globals. */
2988 for (psy = psyms; (sy = *psy) != NULL; ++psy)
2990 asection *s;
2992 s = psecs[sy - syms];
2994 /* Global syms might be improperly typed functions. */
2995 if (ELF_ST_TYPE (sy->st_info) != STT_FUNC
2996 && ELF_ST_BIND (sy->st_info) == STB_GLOBAL)
2998 if (!maybe_insert_function (s, sy, FALSE, FALSE))
2999 return FALSE;
3004 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
3006 extern const bfd_target bfd_elf32_spu_vec;
3007 asection *sec;
3009 if (ibfd->xvec != &bfd_elf32_spu_vec)
3010 continue;
3012 /* Some of the symbols we've installed as marking the
3013 beginning of functions may have a size of zero. Extend
3014 the range of such functions to the beginning of the
3015 next symbol of interest. */
3016 for (sec = ibfd->sections; sec != NULL; sec = sec->next)
3017 if (interesting_section (sec))
3019 struct _spu_elf_section_data *sec_data;
3020 struct spu_elf_stack_info *sinfo;
3022 sec_data = spu_elf_section_data (sec);
3023 sinfo = sec_data->u.i.stack_info;
3024 if (sinfo != NULL && sinfo->num_fun != 0)
3026 int fun_idx;
3027 bfd_vma hi = sec->size;
3029 for (fun_idx = sinfo->num_fun; --fun_idx >= 0; )
3031 sinfo->fun[fun_idx].hi = hi;
3032 hi = sinfo->fun[fun_idx].lo;
3035 sinfo->fun[0].lo = 0;
3037 /* No symbols in this section. Must be .init or .fini
3038 or something similar. */
3039 else if (!pasted_function (sec))
3040 return FALSE;
3045 for (ibfd = info->input_bfds, bfd_idx = 0;
3046 ibfd != NULL;
3047 ibfd = ibfd->link_next, bfd_idx++)
3049 if (psym_arr[bfd_idx] == NULL)
3050 continue;
3052 free (psym_arr[bfd_idx]);
3053 free (sec_arr[bfd_idx]);
3056 free (psym_arr);
3057 free (sec_arr);
3059 return TRUE;
3062 /* Iterate over all function_info we have collected, calling DOIT on
3063 each node if ROOT_ONLY is false. Only call DOIT on root nodes
3064 if ROOT_ONLY. */
3066 static bfd_boolean
3067 for_each_node (bfd_boolean (*doit) (struct function_info *,
3068 struct bfd_link_info *,
3069 void *),
3070 struct bfd_link_info *info,
3071 void *param,
3072 int root_only)
3074 bfd *ibfd;
3076 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
3078 extern const bfd_target bfd_elf32_spu_vec;
3079 asection *sec;
3081 if (ibfd->xvec != &bfd_elf32_spu_vec)
3082 continue;
3084 for (sec = ibfd->sections; sec != NULL; sec = sec->next)
3086 struct _spu_elf_section_data *sec_data;
3087 struct spu_elf_stack_info *sinfo;
3089 if ((sec_data = spu_elf_section_data (sec)) != NULL
3090 && (sinfo = sec_data->u.i.stack_info) != NULL)
3092 int i;
3093 for (i = 0; i < sinfo->num_fun; ++i)
3094 if (!root_only || !sinfo->fun[i].non_root)
3095 if (!doit (&sinfo->fun[i], info, param))
3096 return FALSE;
3100 return TRUE;
3103 /* Transfer call info attached to struct function_info entries for
3104 all of a given function's sections to the first entry. */
3106 static bfd_boolean
3107 transfer_calls (struct function_info *fun,
3108 struct bfd_link_info *info ATTRIBUTE_UNUSED,
3109 void *param ATTRIBUTE_UNUSED)
3111 struct function_info *start = fun->start;
3113 if (start != NULL)
3115 struct call_info *call, *call_next;
3117 while (start->start != NULL)
3118 start = start->start;
3119 for (call = fun->call_list; call != NULL; call = call_next)
3121 call_next = call->next;
3122 if (!insert_callee (start, call))
3123 free (call);
3125 fun->call_list = NULL;
3127 return TRUE;
3130 /* Mark nodes in the call graph that are called by some other node. */
3132 static bfd_boolean
3133 mark_non_root (struct function_info *fun,
3134 struct bfd_link_info *info ATTRIBUTE_UNUSED,
3135 void *param ATTRIBUTE_UNUSED)
3137 struct call_info *call;
3139 if (fun->visit1)
3140 return TRUE;
3141 fun->visit1 = TRUE;
3142 for (call = fun->call_list; call; call = call->next)
3144 call->fun->non_root = TRUE;
3145 mark_non_root (call->fun, 0, 0);
3147 return TRUE;
3150 /* Remove cycles from the call graph. Set depth of nodes. */
3152 static bfd_boolean
3153 remove_cycles (struct function_info *fun,
3154 struct bfd_link_info *info,
3155 void *param)
3157 struct call_info **callp, *call;
3158 unsigned int depth = *(unsigned int *) param;
3159 unsigned int max_depth = depth;
3161 fun->depth = depth;
3162 fun->visit2 = TRUE;
3163 fun->marking = TRUE;
3165 callp = &fun->call_list;
3166 while ((call = *callp) != NULL)
3168 call->max_depth = depth + !call->is_pasted;
3169 if (!call->fun->visit2)
3171 if (!remove_cycles (call->fun, info, &call->max_depth))
3172 return FALSE;
3173 if (max_depth < call->max_depth)
3174 max_depth = call->max_depth;
3176 else if (call->fun->marking)
3178 struct spu_link_hash_table *htab = spu_hash_table (info);
3180 if (!htab->params->auto_overlay
3181 && htab->params->stack_analysis)
3183 const char *f1 = func_name (fun);
3184 const char *f2 = func_name (call->fun);
3186 info->callbacks->info (_("Stack analysis will ignore the call "
3187 "from %s to %s\n"),
3188 f1, f2);
3190 *callp = call->next;
3191 free (call);
3192 continue;
3194 callp = &call->next;
3196 fun->marking = FALSE;
3197 *(unsigned int *) param = max_depth;
3198 return TRUE;
3201 /* Check that we actually visited all nodes in remove_cycles. If we
3202 didn't, then there is some cycle in the call graph not attached to
3203 any root node. Arbitrarily choose a node in the cycle as a new
3204 root and break the cycle. */
3206 static bfd_boolean
3207 mark_detached_root (struct function_info *fun,
3208 struct bfd_link_info *info,
3209 void *param)
3211 if (fun->visit2)
3212 return TRUE;
3213 fun->non_root = FALSE;
3214 *(unsigned int *) param = 0;
3215 return remove_cycles (fun, info, param);
3218 /* Populate call_list for each function. */
3220 static bfd_boolean
3221 build_call_tree (struct bfd_link_info *info)
3223 bfd *ibfd;
3224 unsigned int depth;
3226 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
3228 extern const bfd_target bfd_elf32_spu_vec;
3229 asection *sec;
3231 if (ibfd->xvec != &bfd_elf32_spu_vec)
3232 continue;
3234 for (sec = ibfd->sections; sec != NULL; sec = sec->next)
3235 if (!mark_functions_via_relocs (sec, info, TRUE))
3236 return FALSE;
3239 /* Transfer call info from hot/cold section part of function
3240 to main entry. */
3241 if (!spu_hash_table (info)->params->auto_overlay
3242 && !for_each_node (transfer_calls, info, 0, FALSE))
3243 return FALSE;
3245 /* Find the call graph root(s). */
3246 if (!for_each_node (mark_non_root, info, 0, FALSE))
3247 return FALSE;
3249 /* Remove cycles from the call graph. We start from the root node(s)
3250 so that we break cycles in a reasonable place. */
3251 depth = 0;
3252 if (!for_each_node (remove_cycles, info, &depth, TRUE))
3253 return FALSE;
3255 return for_each_node (mark_detached_root, info, &depth, FALSE);
3258 /* qsort predicate to sort calls by priority, max_depth then count. */
3260 static int
3261 sort_calls (const void *a, const void *b)
3263 struct call_info *const *c1 = a;
3264 struct call_info *const *c2 = b;
3265 int delta;
3267 delta = (*c2)->priority - (*c1)->priority;
3268 if (delta != 0)
3269 return delta;
3271 delta = (*c2)->max_depth - (*c1)->max_depth;
3272 if (delta != 0)
3273 return delta;
3275 delta = (*c2)->count - (*c1)->count;
3276 if (delta != 0)
3277 return delta;
3279 return (char *) c1 - (char *) c2;
3282 struct _mos_param {
3283 unsigned int max_overlay_size;
3286 /* Set linker_mark and gc_mark on any sections that we will put in
3287 overlays. These flags are used by the generic ELF linker, but we
3288 won't be continuing on to bfd_elf_final_link so it is OK to use
3289 them. linker_mark is clear before we get here. Set segment_mark
3290 on sections that are part of a pasted function (excluding the last
3291 section).
3293 Set up function rodata section if --overlay-rodata. We don't
3294 currently include merged string constant rodata sections since
3296 Sort the call graph so that the deepest nodes will be visited
3297 first. */
3299 static bfd_boolean
3300 mark_overlay_section (struct function_info *fun,
3301 struct bfd_link_info *info,
3302 void *param)
3304 struct call_info *call;
3305 unsigned int count;
3306 struct _mos_param *mos_param = param;
3307 struct spu_link_hash_table *htab = spu_hash_table (info);
3309 if (fun->visit4)
3310 return TRUE;
3312 fun->visit4 = TRUE;
3313 if (!fun->sec->linker_mark
3314 && (htab->params->ovly_flavour != ovly_soft_icache
3315 || htab->params->non_ia_text
3316 || strncmp (fun->sec->name, ".text.ia.", 9) == 0))
3318 unsigned int size;
3320 fun->sec->linker_mark = 1;
3321 fun->sec->gc_mark = 1;
3322 fun->sec->segment_mark = 0;
3323 /* Ensure SEC_CODE is set on this text section (it ought to
3324 be!), and SEC_CODE is clear on rodata sections. We use
3325 this flag to differentiate the two overlay section types. */
3326 fun->sec->flags |= SEC_CODE;
3328 size = fun->sec->size;
3329 if (htab->params->auto_overlay & OVERLAY_RODATA)
3331 char *name = NULL;
3333 /* Find the rodata section corresponding to this function's
3334 text section. */
3335 if (strcmp (fun->sec->name, ".text") == 0)
3337 name = bfd_malloc (sizeof (".rodata"));
3338 if (name == NULL)
3339 return FALSE;
3340 memcpy (name, ".rodata", sizeof (".rodata"));
3342 else if (strncmp (fun->sec->name, ".text.", 6) == 0)
3344 size_t len = strlen (fun->sec->name);
3345 name = bfd_malloc (len + 3);
3346 if (name == NULL)
3347 return FALSE;
3348 memcpy (name, ".rodata", sizeof (".rodata"));
3349 memcpy (name + 7, fun->sec->name + 5, len - 4);
3351 else if (strncmp (fun->sec->name, ".gnu.linkonce.t.", 16) == 0)
3353 size_t len = strlen (fun->sec->name) + 1;
3354 name = bfd_malloc (len);
3355 if (name == NULL)
3356 return FALSE;
3357 memcpy (name, fun->sec->name, len);
3358 name[14] = 'r';
3361 if (name != NULL)
3363 asection *rodata = NULL;
3364 asection *group_sec = elf_section_data (fun->sec)->next_in_group;
3365 if (group_sec == NULL)
3366 rodata = bfd_get_section_by_name (fun->sec->owner, name);
3367 else
3368 while (group_sec != NULL && group_sec != fun->sec)
3370 if (strcmp (group_sec->name, name) == 0)
3372 rodata = group_sec;
3373 break;
3375 group_sec = elf_section_data (group_sec)->next_in_group;
3377 fun->rodata = rodata;
3378 if (fun->rodata)
3380 size += fun->rodata->size;
3381 if (htab->params->line_size != 0
3382 && size > htab->params->line_size)
3384 size -= fun->rodata->size;
3385 fun->rodata = NULL;
3387 else
3389 fun->rodata->linker_mark = 1;
3390 fun->rodata->gc_mark = 1;
3391 fun->rodata->flags &= ~SEC_CODE;
3394 free (name);
3397 if (mos_param->max_overlay_size < size)
3398 mos_param->max_overlay_size = size;
3401 for (count = 0, call = fun->call_list; call != NULL; call = call->next)
3402 count += 1;
3404 if (count > 1)
3406 struct call_info **calls = bfd_malloc (count * sizeof (*calls));
3407 if (calls == NULL)
3408 return FALSE;
3410 for (count = 0, call = fun->call_list; call != NULL; call = call->next)
3411 calls[count++] = call;
3413 qsort (calls, count, sizeof (*calls), sort_calls);
3415 fun->call_list = NULL;
3416 while (count != 0)
3418 --count;
3419 calls[count]->next = fun->call_list;
3420 fun->call_list = calls[count];
3422 free (calls);
3425 for (call = fun->call_list; call != NULL; call = call->next)
3427 if (call->is_pasted)
3429 /* There can only be one is_pasted call per function_info. */
3430 BFD_ASSERT (!fun->sec->segment_mark);
3431 fun->sec->segment_mark = 1;
3433 if (!mark_overlay_section (call->fun, info, param))
3434 return FALSE;
3437 /* Don't put entry code into an overlay. The overlay manager needs
3438 a stack! Also, don't mark .ovl.init as an overlay. */
3439 if (fun->lo + fun->sec->output_offset + fun->sec->output_section->vma
3440 == info->output_bfd->start_address
3441 || strncmp (fun->sec->output_section->name, ".ovl.init", 9) == 0)
3443 fun->sec->linker_mark = 0;
3444 if (fun->rodata != NULL)
3445 fun->rodata->linker_mark = 0;
3447 return TRUE;
3450 /* If non-zero then unmark functions called from those within sections
3451 that we need to unmark. Unfortunately this isn't reliable since the
3452 call graph cannot know the destination of function pointer calls. */
3453 #define RECURSE_UNMARK 0
3455 struct _uos_param {
3456 asection *exclude_input_section;
3457 asection *exclude_output_section;
3458 unsigned long clearing;
3461 /* Undo some of mark_overlay_section's work. */
3463 static bfd_boolean
3464 unmark_overlay_section (struct function_info *fun,
3465 struct bfd_link_info *info,
3466 void *param)
3468 struct call_info *call;
3469 struct _uos_param *uos_param = param;
3470 unsigned int excluded = 0;
3472 if (fun->visit5)
3473 return TRUE;
3475 fun->visit5 = TRUE;
3477 excluded = 0;
3478 if (fun->sec == uos_param->exclude_input_section
3479 || fun->sec->output_section == uos_param->exclude_output_section)
3480 excluded = 1;
3482 if (RECURSE_UNMARK)
3483 uos_param->clearing += excluded;
3485 if (RECURSE_UNMARK ? uos_param->clearing : excluded)
3487 fun->sec->linker_mark = 0;
3488 if (fun->rodata)
3489 fun->rodata->linker_mark = 0;
3492 for (call = fun->call_list; call != NULL; call = call->next)
3493 if (!unmark_overlay_section (call->fun, info, param))
3494 return FALSE;
3496 if (RECURSE_UNMARK)
3497 uos_param->clearing -= excluded;
3498 return TRUE;
3501 struct _cl_param {
3502 unsigned int lib_size;
3503 asection **lib_sections;
3506 /* Add sections we have marked as belonging to overlays to an array
3507 for consideration as non-overlay sections. The array consist of
3508 pairs of sections, (text,rodata), for functions in the call graph. */
3510 static bfd_boolean
3511 collect_lib_sections (struct function_info *fun,
3512 struct bfd_link_info *info,
3513 void *param)
3515 struct _cl_param *lib_param = param;
3516 struct call_info *call;
3517 unsigned int size;
3519 if (fun->visit6)
3520 return TRUE;
3522 fun->visit6 = TRUE;
3523 if (!fun->sec->linker_mark || !fun->sec->gc_mark || fun->sec->segment_mark)
3524 return TRUE;
3526 size = fun->sec->size;
3527 if (fun->rodata)
3528 size += fun->rodata->size;
3530 if (size <= lib_param->lib_size)
3532 *lib_param->lib_sections++ = fun->sec;
3533 fun->sec->gc_mark = 0;
3534 if (fun->rodata && fun->rodata->linker_mark && fun->rodata->gc_mark)
3536 *lib_param->lib_sections++ = fun->rodata;
3537 fun->rodata->gc_mark = 0;
3539 else
3540 *lib_param->lib_sections++ = NULL;
3543 for (call = fun->call_list; call != NULL; call = call->next)
3544 collect_lib_sections (call->fun, info, param);
3546 return TRUE;
3549 /* qsort predicate to sort sections by call count. */
3551 static int
3552 sort_lib (const void *a, const void *b)
3554 asection *const *s1 = a;
3555 asection *const *s2 = b;
3556 struct _spu_elf_section_data *sec_data;
3557 struct spu_elf_stack_info *sinfo;
3558 int delta;
3560 delta = 0;
3561 if ((sec_data = spu_elf_section_data (*s1)) != NULL
3562 && (sinfo = sec_data->u.i.stack_info) != NULL)
3564 int i;
3565 for (i = 0; i < sinfo->num_fun; ++i)
3566 delta -= sinfo->fun[i].call_count;
3569 if ((sec_data = spu_elf_section_data (*s2)) != NULL
3570 && (sinfo = sec_data->u.i.stack_info) != NULL)
3572 int i;
3573 for (i = 0; i < sinfo->num_fun; ++i)
3574 delta += sinfo->fun[i].call_count;
3577 if (delta != 0)
3578 return delta;
3580 return s1 - s2;
3583 /* Remove some sections from those marked to be in overlays. Choose
3584 those that are called from many places, likely library functions. */
3586 static unsigned int
3587 auto_ovl_lib_functions (struct bfd_link_info *info, unsigned int lib_size)
3589 bfd *ibfd;
3590 asection **lib_sections;
3591 unsigned int i, lib_count;
3592 struct _cl_param collect_lib_param;
3593 struct function_info dummy_caller;
3594 struct spu_link_hash_table *htab;
3596 memset (&dummy_caller, 0, sizeof (dummy_caller));
3597 lib_count = 0;
3598 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
3600 extern const bfd_target bfd_elf32_spu_vec;
3601 asection *sec;
3603 if (ibfd->xvec != &bfd_elf32_spu_vec)
3604 continue;
3606 for (sec = ibfd->sections; sec != NULL; sec = sec->next)
3607 if (sec->linker_mark
3608 && sec->size < lib_size
3609 && (sec->flags & SEC_CODE) != 0)
3610 lib_count += 1;
3612 lib_sections = bfd_malloc (lib_count * 2 * sizeof (*lib_sections));
3613 if (lib_sections == NULL)
3614 return (unsigned int) -1;
3615 collect_lib_param.lib_size = lib_size;
3616 collect_lib_param.lib_sections = lib_sections;
3617 if (!for_each_node (collect_lib_sections, info, &collect_lib_param,
3618 TRUE))
3619 return (unsigned int) -1;
3620 lib_count = (collect_lib_param.lib_sections - lib_sections) / 2;
3622 /* Sort sections so that those with the most calls are first. */
3623 if (lib_count > 1)
3624 qsort (lib_sections, lib_count, 2 * sizeof (*lib_sections), sort_lib);
3626 htab = spu_hash_table (info);
3627 for (i = 0; i < lib_count; i++)
3629 unsigned int tmp, stub_size;
3630 asection *sec;
3631 struct _spu_elf_section_data *sec_data;
3632 struct spu_elf_stack_info *sinfo;
3634 sec = lib_sections[2 * i];
3635 /* If this section is OK, its size must be less than lib_size. */
3636 tmp = sec->size;
3637 /* If it has a rodata section, then add that too. */
3638 if (lib_sections[2 * i + 1])
3639 tmp += lib_sections[2 * i + 1]->size;
3640 /* Add any new overlay call stubs needed by the section. */
3641 stub_size = 0;
3642 if (tmp < lib_size
3643 && (sec_data = spu_elf_section_data (sec)) != NULL
3644 && (sinfo = sec_data->u.i.stack_info) != NULL)
3646 int k;
3647 struct call_info *call;
3649 for (k = 0; k < sinfo->num_fun; ++k)
3650 for (call = sinfo->fun[k].call_list; call; call = call->next)
3651 if (call->fun->sec->linker_mark)
3653 struct call_info *p;
3654 for (p = dummy_caller.call_list; p; p = p->next)
3655 if (p->fun == call->fun)
3656 break;
3657 if (!p)
3658 stub_size += ovl_stub_size (htab->params->ovly_flavour);
3661 if (tmp + stub_size < lib_size)
3663 struct call_info **pp, *p;
3665 /* This section fits. Mark it as non-overlay. */
3666 lib_sections[2 * i]->linker_mark = 0;
3667 if (lib_sections[2 * i + 1])
3668 lib_sections[2 * i + 1]->linker_mark = 0;
3669 lib_size -= tmp + stub_size;
3670 /* Call stubs to the section we just added are no longer
3671 needed. */
3672 pp = &dummy_caller.call_list;
3673 while ((p = *pp) != NULL)
3674 if (!p->fun->sec->linker_mark)
3676 lib_size += ovl_stub_size (htab->params->ovly_flavour);
3677 *pp = p->next;
3678 free (p);
3680 else
3681 pp = &p->next;
3682 /* Add new call stubs to dummy_caller. */
3683 if ((sec_data = spu_elf_section_data (sec)) != NULL
3684 && (sinfo = sec_data->u.i.stack_info) != NULL)
3686 int k;
3687 struct call_info *call;
3689 for (k = 0; k < sinfo->num_fun; ++k)
3690 for (call = sinfo->fun[k].call_list;
3691 call;
3692 call = call->next)
3693 if (call->fun->sec->linker_mark)
3695 struct call_info *callee;
3696 callee = bfd_malloc (sizeof (*callee));
3697 if (callee == NULL)
3698 return (unsigned int) -1;
3699 *callee = *call;
3700 if (!insert_callee (&dummy_caller, callee))
3701 free (callee);
3706 while (dummy_caller.call_list != NULL)
3708 struct call_info *call = dummy_caller.call_list;
3709 dummy_caller.call_list = call->next;
3710 free (call);
3712 for (i = 0; i < 2 * lib_count; i++)
3713 if (lib_sections[i])
3714 lib_sections[i]->gc_mark = 1;
3715 free (lib_sections);
3716 return lib_size;
3719 /* Build an array of overlay sections. The deepest node's section is
3720 added first, then its parent node's section, then everything called
3721 from the parent section. The idea being to group sections to
3722 minimise calls between different overlays. */
3724 static bfd_boolean
3725 collect_overlays (struct function_info *fun,
3726 struct bfd_link_info *info,
3727 void *param)
3729 struct call_info *call;
3730 bfd_boolean added_fun;
3731 asection ***ovly_sections = param;
3733 if (fun->visit7)
3734 return TRUE;
3736 fun->visit7 = TRUE;
3737 for (call = fun->call_list; call != NULL; call = call->next)
3738 if (!call->is_pasted)
3740 if (!collect_overlays (call->fun, info, ovly_sections))
3741 return FALSE;
3742 break;
3745 added_fun = FALSE;
3746 if (fun->sec->linker_mark && fun->sec->gc_mark)
3748 fun->sec->gc_mark = 0;
3749 *(*ovly_sections)++ = fun->sec;
3750 if (fun->rodata && fun->rodata->linker_mark && fun->rodata->gc_mark)
3752 fun->rodata->gc_mark = 0;
3753 *(*ovly_sections)++ = fun->rodata;
3755 else
3756 *(*ovly_sections)++ = NULL;
3757 added_fun = TRUE;
3759 /* Pasted sections must stay with the first section. We don't
3760 put pasted sections in the array, just the first section.
3761 Mark subsequent sections as already considered. */
3762 if (fun->sec->segment_mark)
3764 struct function_info *call_fun = fun;
3767 for (call = call_fun->call_list; call != NULL; call = call->next)
3768 if (call->is_pasted)
3770 call_fun = call->fun;
3771 call_fun->sec->gc_mark = 0;
3772 if (call_fun->rodata)
3773 call_fun->rodata->gc_mark = 0;
3774 break;
3776 if (call == NULL)
3777 abort ();
3779 while (call_fun->sec->segment_mark);
3783 for (call = fun->call_list; call != NULL; call = call->next)
3784 if (!collect_overlays (call->fun, info, ovly_sections))
3785 return FALSE;
3787 if (added_fun)
3789 struct _spu_elf_section_data *sec_data;
3790 struct spu_elf_stack_info *sinfo;
3792 if ((sec_data = spu_elf_section_data (fun->sec)) != NULL
3793 && (sinfo = sec_data->u.i.stack_info) != NULL)
3795 int i;
3796 for (i = 0; i < sinfo->num_fun; ++i)
3797 if (!collect_overlays (&sinfo->fun[i], info, ovly_sections))
3798 return FALSE;
3802 return TRUE;
3805 struct _sum_stack_param {
3806 size_t cum_stack;
3807 size_t overall_stack;
3808 bfd_boolean emit_stack_syms;
3811 /* Descend the call graph for FUN, accumulating total stack required. */
3813 static bfd_boolean
3814 sum_stack (struct function_info *fun,
3815 struct bfd_link_info *info,
3816 void *param)
3818 struct call_info *call;
3819 struct function_info *max;
3820 size_t stack, cum_stack;
3821 const char *f1;
3822 bfd_boolean has_call;
3823 struct _sum_stack_param *sum_stack_param = param;
3824 struct spu_link_hash_table *htab;
3826 cum_stack = fun->stack;
3827 sum_stack_param->cum_stack = cum_stack;
3828 if (fun->visit3)
3829 return TRUE;
3831 has_call = FALSE;
3832 max = NULL;
3833 for (call = fun->call_list; call; call = call->next)
3835 if (!call->is_pasted)
3836 has_call = TRUE;
3837 if (!sum_stack (call->fun, info, sum_stack_param))
3838 return FALSE;
3839 stack = sum_stack_param->cum_stack;
3840 /* Include caller stack for normal calls, don't do so for
3841 tail calls. fun->stack here is local stack usage for
3842 this function. */
3843 if (!call->is_tail || call->is_pasted || call->fun->start != NULL)
3844 stack += fun->stack;
3845 if (cum_stack < stack)
3847 cum_stack = stack;
3848 max = call->fun;
3852 sum_stack_param->cum_stack = cum_stack;
3853 stack = fun->stack;
3854 /* Now fun->stack holds cumulative stack. */
3855 fun->stack = cum_stack;
3856 fun->visit3 = TRUE;
3858 if (!fun->non_root
3859 && sum_stack_param->overall_stack < cum_stack)
3860 sum_stack_param->overall_stack = cum_stack;
3862 htab = spu_hash_table (info);
3863 if (htab->params->auto_overlay)
3864 return TRUE;
3866 f1 = func_name (fun);
3867 if (htab->params->stack_analysis)
3869 if (!fun->non_root)
3870 info->callbacks->info (_(" %s: 0x%v\n"), f1, (bfd_vma) cum_stack);
3871 info->callbacks->minfo (_("%s: 0x%v 0x%v\n"),
3872 f1, (bfd_vma) stack, (bfd_vma) cum_stack);
3874 if (has_call)
3876 info->callbacks->minfo (_(" calls:\n"));
3877 for (call = fun->call_list; call; call = call->next)
3878 if (!call->is_pasted)
3880 const char *f2 = func_name (call->fun);
3881 const char *ann1 = call->fun == max ? "*" : " ";
3882 const char *ann2 = call->is_tail ? "t" : " ";
3884 info->callbacks->minfo (_(" %s%s %s\n"), ann1, ann2, f2);
3889 if (sum_stack_param->emit_stack_syms)
3891 char *name = bfd_malloc (18 + strlen (f1));
3892 struct elf_link_hash_entry *h;
3894 if (name == NULL)
3895 return FALSE;
3897 if (fun->global || ELF_ST_BIND (fun->u.sym->st_info) == STB_GLOBAL)
3898 sprintf (name, "__stack_%s", f1);
3899 else
3900 sprintf (name, "__stack_%x_%s", fun->sec->id & 0xffffffff, f1);
3902 h = elf_link_hash_lookup (&htab->elf, name, TRUE, TRUE, FALSE);
3903 free (name);
3904 if (h != NULL
3905 && (h->root.type == bfd_link_hash_new
3906 || h->root.type == bfd_link_hash_undefined
3907 || h->root.type == bfd_link_hash_undefweak))
3909 h->root.type = bfd_link_hash_defined;
3910 h->root.u.def.section = bfd_abs_section_ptr;
3911 h->root.u.def.value = cum_stack;
3912 h->size = 0;
3913 h->type = 0;
3914 h->ref_regular = 1;
3915 h->def_regular = 1;
3916 h->ref_regular_nonweak = 1;
3917 h->forced_local = 1;
3918 h->non_elf = 0;
3922 return TRUE;
3925 /* SEC is part of a pasted function. Return the call_info for the
3926 next section of this function. */
3928 static struct call_info *
3929 find_pasted_call (asection *sec)
3931 struct _spu_elf_section_data *sec_data = spu_elf_section_data (sec);
3932 struct spu_elf_stack_info *sinfo = sec_data->u.i.stack_info;
3933 struct call_info *call;
3934 int k;
3936 for (k = 0; k < sinfo->num_fun; ++k)
3937 for (call = sinfo->fun[k].call_list; call != NULL; call = call->next)
3938 if (call->is_pasted)
3939 return call;
3940 abort ();
3941 return 0;
3944 /* qsort predicate to sort bfds by file name. */
3946 static int
3947 sort_bfds (const void *a, const void *b)
3949 bfd *const *abfd1 = a;
3950 bfd *const *abfd2 = b;
3952 return strcmp ((*abfd1)->filename, (*abfd2)->filename);
3955 static unsigned int
3956 print_one_overlay_section (FILE *script,
3957 unsigned int base,
3958 unsigned int count,
3959 unsigned int ovlynum,
3960 unsigned int *ovly_map,
3961 asection **ovly_sections,
3962 struct bfd_link_info *info)
3964 unsigned int j;
3966 for (j = base; j < count && ovly_map[j] == ovlynum; j++)
3968 asection *sec = ovly_sections[2 * j];
3970 if (fprintf (script, " %s%c%s (%s)\n",
3971 (sec->owner->my_archive != NULL
3972 ? sec->owner->my_archive->filename : ""),
3973 info->path_separator,
3974 sec->owner->filename,
3975 sec->name) <= 0)
3976 return -1;
3977 if (sec->segment_mark)
3979 struct call_info *call = find_pasted_call (sec);
3980 while (call != NULL)
3982 struct function_info *call_fun = call->fun;
3983 sec = call_fun->sec;
3984 if (fprintf (script, " %s%c%s (%s)\n",
3985 (sec->owner->my_archive != NULL
3986 ? sec->owner->my_archive->filename : ""),
3987 info->path_separator,
3988 sec->owner->filename,
3989 sec->name) <= 0)
3990 return -1;
3991 for (call = call_fun->call_list; call; call = call->next)
3992 if (call->is_pasted)
3993 break;
3998 for (j = base; j < count && ovly_map[j] == ovlynum; j++)
4000 asection *sec = ovly_sections[2 * j + 1];
4001 if (sec != NULL
4002 && fprintf (script, " %s%c%s (%s)\n",
4003 (sec->owner->my_archive != NULL
4004 ? sec->owner->my_archive->filename : ""),
4005 info->path_separator,
4006 sec->owner->filename,
4007 sec->name) <= 0)
4008 return -1;
4010 sec = ovly_sections[2 * j];
4011 if (sec->segment_mark)
4013 struct call_info *call = find_pasted_call (sec);
4014 while (call != NULL)
4016 struct function_info *call_fun = call->fun;
4017 sec = call_fun->rodata;
4018 if (sec != NULL
4019 && fprintf (script, " %s%c%s (%s)\n",
4020 (sec->owner->my_archive != NULL
4021 ? sec->owner->my_archive->filename : ""),
4022 info->path_separator,
4023 sec->owner->filename,
4024 sec->name) <= 0)
4025 return -1;
4026 for (call = call_fun->call_list; call; call = call->next)
4027 if (call->is_pasted)
4028 break;
4033 return j;
4036 /* Handle --auto-overlay. */
4038 static void spu_elf_auto_overlay (struct bfd_link_info *)
4039 ATTRIBUTE_NORETURN;
4041 static void
4042 spu_elf_auto_overlay (struct bfd_link_info *info)
4044 bfd *ibfd;
4045 bfd **bfd_arr;
4046 struct elf_segment_map *m;
4047 unsigned int fixed_size, lo, hi;
4048 struct spu_link_hash_table *htab;
4049 unsigned int base, i, count, bfd_count;
4050 unsigned int region, ovlynum;
4051 asection **ovly_sections, **ovly_p;
4052 unsigned int *ovly_map;
4053 FILE *script;
4054 unsigned int total_overlay_size, overlay_size;
4055 const char *ovly_mgr_entry;
4056 struct elf_link_hash_entry *h;
4057 struct _mos_param mos_param;
4058 struct _uos_param uos_param;
4059 struct function_info dummy_caller;
4061 /* Find the extents of our loadable image. */
4062 lo = (unsigned int) -1;
4063 hi = 0;
4064 for (m = elf_tdata (info->output_bfd)->segment_map; m != NULL; m = m->next)
4065 if (m->p_type == PT_LOAD)
4066 for (i = 0; i < m->count; i++)
4067 if (m->sections[i]->size != 0)
4069 if (m->sections[i]->vma < lo)
4070 lo = m->sections[i]->vma;
4071 if (m->sections[i]->vma + m->sections[i]->size - 1 > hi)
4072 hi = m->sections[i]->vma + m->sections[i]->size - 1;
4074 fixed_size = hi + 1 - lo;
4076 if (!discover_functions (info))
4077 goto err_exit;
4079 if (!build_call_tree (info))
4080 goto err_exit;
4082 uos_param.exclude_input_section = 0;
4083 uos_param.exclude_output_section
4084 = bfd_get_section_by_name (info->output_bfd, ".interrupt");
4086 htab = spu_hash_table (info);
4087 ovly_mgr_entry = "__ovly_load";
4088 if (htab->params->ovly_flavour == ovly_soft_icache)
4089 ovly_mgr_entry = "__icache_br_handler";
4090 h = elf_link_hash_lookup (&htab->elf, ovly_mgr_entry,
4091 FALSE, FALSE, FALSE);
4092 if (h != NULL
4093 && (h->root.type == bfd_link_hash_defined
4094 || h->root.type == bfd_link_hash_defweak)
4095 && h->def_regular)
4097 /* We have a user supplied overlay manager. */
4098 uos_param.exclude_input_section = h->root.u.def.section;
4100 else
4102 /* If no user overlay manager, spu_elf_load_ovl_mgr will add our
4103 builtin version to .text, and will adjust .text size. */
4104 fixed_size += (*htab->params->spu_elf_load_ovl_mgr) ();
4107 /* Mark overlay sections, and find max overlay section size. */
4108 mos_param.max_overlay_size = 0;
4109 if (!for_each_node (mark_overlay_section, info, &mos_param, TRUE))
4110 goto err_exit;
4112 /* We can't put the overlay manager or interrupt routines in
4113 overlays. */
4114 uos_param.clearing = 0;
4115 if ((uos_param.exclude_input_section
4116 || uos_param.exclude_output_section)
4117 && !for_each_node (unmark_overlay_section, info, &uos_param, TRUE))
4118 goto err_exit;
4120 bfd_count = 0;
4121 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
4122 ++bfd_count;
4123 bfd_arr = bfd_malloc (bfd_count * sizeof (*bfd_arr));
4124 if (bfd_arr == NULL)
4125 goto err_exit;
4127 /* Count overlay sections, and subtract their sizes from "fixed_size". */
4128 count = 0;
4129 bfd_count = 0;
4130 total_overlay_size = 0;
4131 for (ibfd = info->input_bfds; ibfd != NULL; ibfd = ibfd->link_next)
4133 extern const bfd_target bfd_elf32_spu_vec;
4134 asection *sec;
4135 unsigned int old_count;
4137 if (ibfd->xvec != &bfd_elf32_spu_vec)
4138 continue;
4140 old_count = count;
4141 for (sec = ibfd->sections; sec != NULL; sec = sec->next)
4142 if (sec->linker_mark)
4144 if ((sec->flags & SEC_CODE) != 0)
4145 count += 1;
4146 fixed_size -= sec->size;
4147 total_overlay_size += sec->size;
4149 else if ((sec->flags & (SEC_ALLOC | SEC_LOAD)) == (SEC_ALLOC | SEC_LOAD)
4150 && sec->output_section->owner == info->output_bfd
4151 && strncmp (sec->output_section->name, ".ovl.init", 9) == 0)
4152 fixed_size -= sec->size;
4153 if (count != old_count)
4154 bfd_arr[bfd_count++] = ibfd;
4157 /* Since the overlay link script selects sections by file name and
4158 section name, ensure that file names are unique. */
4159 if (bfd_count > 1)
4161 bfd_boolean ok = TRUE;
4163 qsort (bfd_arr, bfd_count, sizeof (*bfd_arr), sort_bfds);
4164 for (i = 1; i < bfd_count; ++i)
4165 if (strcmp (bfd_arr[i - 1]->filename, bfd_arr[i]->filename) == 0)
4167 if (bfd_arr[i - 1]->my_archive == bfd_arr[i]->my_archive)
4169 if (bfd_arr[i - 1]->my_archive && bfd_arr[i]->my_archive)
4170 info->callbacks->einfo (_("%s duplicated in %s\n"),
4171 bfd_arr[i]->filename,
4172 bfd_arr[i]->my_archive->filename);
4173 else
4174 info->callbacks->einfo (_("%s duplicated\n"),
4175 bfd_arr[i]->filename);
4176 ok = FALSE;
4179 if (!ok)
4181 info->callbacks->einfo (_("sorry, no support for duplicate "
4182 "object files in auto-overlay script\n"));
4183 bfd_set_error (bfd_error_bad_value);
4184 goto err_exit;
4187 free (bfd_arr);
4189 if (htab->reserved == 0)
4191 struct _sum_stack_param sum_stack_param;
4193 sum_stack_param.emit_stack_syms = 0;
4194 sum_stack_param.overall_stack = 0;
4195 if (!for_each_node (sum_stack, info, &sum_stack_param, TRUE))
4196 goto err_exit;
4197 htab->reserved = sum_stack_param.overall_stack + htab->extra_stack_space;
4199 fixed_size += htab->reserved;
4200 fixed_size += htab->non_ovly_stub * ovl_stub_size (htab->params->ovly_flavour);
4201 if (fixed_size + mos_param.max_overlay_size <= htab->local_store)
4203 if (htab->params->ovly_flavour == ovly_soft_icache)
4205 /* Stubs in the non-icache area are bigger. */
4206 fixed_size += htab->non_ovly_stub * 16;
4207 /* Space for icache manager tables.
4208 a) Tag array, one quadword per cache line.
4209 - word 0: ia address of present line, init to zero.
4210 - word 1: link locator. link_elem=stub_addr/2+locator
4211 - halfwords 4-7: head/tail pointers for linked lists. */
4212 fixed_size += 16 << htab->num_lines_log2;
4213 /* b) Linked list elements, max_branch per line. */
4214 fixed_size += htab->params->max_branch << (htab->num_lines_log2 + 4);
4215 /* c) Indirect branch descriptors, 8 quadwords. */
4216 fixed_size += 8 * 16;
4217 /* d) Pointers to __ea backing store, 16 quadwords. */
4218 fixed_size += 16 * 16;
4220 else
4222 /* Guess number of overlays. Assuming overlay buffer is on
4223 average only half full should be conservative. */
4224 ovlynum = (total_overlay_size * 2 * htab->params->num_lines
4225 / (htab->local_store - fixed_size));
4226 /* Space for _ovly_table[], _ovly_buf_table[] and toe. */
4227 fixed_size += ovlynum * 16 + 16 + 4 + 16;
4231 if (fixed_size + mos_param.max_overlay_size > htab->local_store)
4232 info->callbacks->einfo (_("non-overlay size of 0x%v plus maximum overlay "
4233 "size of 0x%v exceeds local store\n"),
4234 (bfd_vma) fixed_size,
4235 (bfd_vma) mos_param.max_overlay_size);
4237 /* Now see if we should put some functions in the non-overlay area. */
4238 else if (fixed_size < htab->overlay_fixed)
4240 unsigned int max_fixed, lib_size;
4242 max_fixed = htab->local_store - mos_param.max_overlay_size;
4243 if (max_fixed > htab->overlay_fixed)
4244 max_fixed = htab->overlay_fixed;
4245 lib_size = max_fixed - fixed_size;
4246 lib_size = auto_ovl_lib_functions (info, lib_size);
4247 if (lib_size == (unsigned int) -1)
4248 goto err_exit;
4249 fixed_size = max_fixed - lib_size;
4252 /* Build an array of sections, suitably sorted to place into
4253 overlays. */
4254 ovly_sections = bfd_malloc (2 * count * sizeof (*ovly_sections));
4255 if (ovly_sections == NULL)
4256 goto err_exit;
4257 ovly_p = ovly_sections;
4258 if (!for_each_node (collect_overlays, info, &ovly_p, TRUE))
4259 goto err_exit;
4260 count = (size_t) (ovly_p - ovly_sections) / 2;
4261 ovly_map = bfd_malloc (count * sizeof (*ovly_map));
4262 if (ovly_map == NULL)
4263 goto err_exit;
4265 memset (&dummy_caller, 0, sizeof (dummy_caller));
4266 overlay_size = (htab->local_store - fixed_size) / htab->params->num_lines;
4267 if (htab->params->line_size != 0)
4268 overlay_size = htab->params->line_size;
4269 base = 0;
4270 ovlynum = 0;
4271 while (base < count)
4273 unsigned int size = 0;
4275 for (i = base; i < count; i++)
4277 asection *sec;
4278 unsigned int tmp;
4279 unsigned int num_stubs;
4280 struct call_info *call, *pasty;
4281 struct _spu_elf_section_data *sec_data;
4282 struct spu_elf_stack_info *sinfo;
4283 int k;
4285 /* See whether we can add this section to the current
4286 overlay without overflowing our overlay buffer. */
4287 sec = ovly_sections[2 * i];
4288 tmp = size + sec->size;
4289 if (ovly_sections[2 * i + 1])
4290 tmp += ovly_sections[2 * i + 1]->size;
4291 if (tmp > overlay_size)
4292 break;
4293 if (sec->segment_mark)
4295 /* Pasted sections must stay together, so add their
4296 sizes too. */
4297 struct call_info *pasty = find_pasted_call (sec);
4298 while (pasty != NULL)
4300 struct function_info *call_fun = pasty->fun;
4301 tmp += call_fun->sec->size;
4302 if (call_fun->rodata)
4303 tmp += call_fun->rodata->size;
4304 for (pasty = call_fun->call_list; pasty; pasty = pasty->next)
4305 if (pasty->is_pasted)
4306 break;
4309 if (tmp > overlay_size)
4310 break;
4312 /* If we add this section, we might need new overlay call
4313 stubs. Add any overlay section calls to dummy_call. */
4314 pasty = NULL;
4315 sec_data = spu_elf_section_data (sec);
4316 sinfo = sec_data->u.i.stack_info;
4317 for (k = 0; k < sinfo->num_fun; ++k)
4318 for (call = sinfo->fun[k].call_list; call; call = call->next)
4319 if (call->is_pasted)
4321 BFD_ASSERT (pasty == NULL);
4322 pasty = call;
4324 else if (call->fun->sec->linker_mark)
4326 if (!copy_callee (&dummy_caller, call))
4327 goto err_exit;
4329 while (pasty != NULL)
4331 struct function_info *call_fun = pasty->fun;
4332 pasty = NULL;
4333 for (call = call_fun->call_list; call; call = call->next)
4334 if (call->is_pasted)
4336 BFD_ASSERT (pasty == NULL);
4337 pasty = call;
4339 else if (!copy_callee (&dummy_caller, call))
4340 goto err_exit;
4343 /* Calculate call stub size. */
4344 num_stubs = 0;
4345 for (call = dummy_caller.call_list; call; call = call->next)
4347 unsigned int k;
4349 ++num_stubs;
4350 /* If the call is within this overlay, we won't need a
4351 stub. */
4352 for (k = base; k < i + 1; k++)
4353 if (call->fun->sec == ovly_sections[2 * k])
4355 --num_stubs;
4356 break;
4359 if (htab->params->ovly_flavour == ovly_soft_icache
4360 && num_stubs > htab->params->max_branch)
4361 break;
4362 if (tmp + num_stubs * ovl_stub_size (htab->params->ovly_flavour)
4363 > overlay_size)
4364 break;
4365 size = tmp;
4368 if (i == base)
4370 info->callbacks->einfo (_("%B:%A%s exceeds overlay size\n"),
4371 ovly_sections[2 * i]->owner,
4372 ovly_sections[2 * i],
4373 ovly_sections[2 * i + 1] ? " + rodata" : "");
4374 bfd_set_error (bfd_error_bad_value);
4375 goto err_exit;
4378 while (dummy_caller.call_list != NULL)
4380 struct call_info *call = dummy_caller.call_list;
4381 dummy_caller.call_list = call->next;
4382 free (call);
4385 ++ovlynum;
4386 while (base < i)
4387 ovly_map[base++] = ovlynum;
4390 script = htab->params->spu_elf_open_overlay_script ();
4392 if (fprintf (script, "SECTIONS\n{\n") <= 0)
4393 goto file_err;
4395 if (htab->params->ovly_flavour == ovly_soft_icache)
4397 if (fprintf (script,
4398 " .data.icache ALIGN (16) : { *(.ovtab) *(.data.icache) }\n"
4399 " . = ALIGN (%u);\n"
4400 " .ovl.init : { *(.ovl.init) }\n"
4401 " . = ABSOLUTE (ADDR (.ovl.init));\n",
4402 htab->params->line_size) <= 0)
4403 goto file_err;
4405 base = 0;
4406 ovlynum = 1;
4407 while (base < count)
4409 unsigned int indx = ovlynum - 1;
4410 unsigned int vma, lma;
4412 vma = (indx & (htab->params->num_lines - 1)) << htab->line_size_log2;
4413 lma = indx << htab->line_size_log2;
4415 if (fprintf (script, " .ovly%u ABSOLUTE (ADDR (.ovl.init)) + %u "
4416 ": AT (ALIGN (LOADADDR (.ovl.init) + SIZEOF (.ovl.init), 16) + %u) {\n",
4417 ovlynum, vma, lma) <= 0)
4418 goto file_err;
4420 base = print_one_overlay_section (script, base, count, ovlynum,
4421 ovly_map, ovly_sections, info);
4422 if (base == (unsigned) -1)
4423 goto file_err;
4425 if (fprintf (script, " }\n") <= 0)
4426 goto file_err;
4428 ovlynum++;
4431 if (fprintf (script, " . = ABSOLUTE (ADDR (.ovl.init)) + %u;\n",
4432 1 << (htab->num_lines_log2 + htab->line_size_log2)) <= 0)
4433 goto file_err;
4435 else
4437 if (fprintf (script,
4438 " . = ALIGN (16);\n"
4439 " .ovl.init : { *(.ovl.init) }\n"
4440 " . = ABSOLUTE (ADDR (.ovl.init));\n") <= 0)
4441 goto file_err;
4443 for (region = 1; region <= htab->params->num_lines; region++)
4445 ovlynum = region;
4446 base = 0;
4447 while (base < count && ovly_map[base] < ovlynum)
4448 base++;
4450 if (base == count)
4451 break;
4453 if (region == 1)
4455 /* We need to set lma since we are overlaying .ovl.init. */
4456 if (fprintf (script,
4457 " OVERLAY : AT (ALIGN (LOADADDR (.ovl.init) + SIZEOF (.ovl.init), 16))\n {\n") <= 0)
4458 goto file_err;
4460 else
4462 if (fprintf (script, " OVERLAY :\n {\n") <= 0)
4463 goto file_err;
4466 while (base < count)
4468 if (fprintf (script, " .ovly%u {\n", ovlynum) <= 0)
4469 goto file_err;
4471 base = print_one_overlay_section (script, base, count, ovlynum,
4472 ovly_map, ovly_sections, info);
4473 if (base == (unsigned) -1)
4474 goto file_err;
4476 if (fprintf (script, " }\n") <= 0)
4477 goto file_err;
4479 ovlynum += htab->params->num_lines;
4480 while (base < count && ovly_map[base] < ovlynum)
4481 base++;
4484 if (fprintf (script, " }\n") <= 0)
4485 goto file_err;
4490 free (ovly_map);
4491 free (ovly_sections);
4493 if (fprintf (script, "}\nINSERT BEFORE .text;\n") <= 0)
4494 goto file_err;
4495 if (fclose (script) != 0)
4496 goto file_err;
4498 if (htab->params->auto_overlay & AUTO_RELINK)
4499 (*htab->params->spu_elf_relink) ();
4501 xexit (0);
4503 file_err:
4504 bfd_set_error (bfd_error_system_call);
4505 err_exit:
4506 info->callbacks->einfo ("%F%P: auto overlay error: %E\n");
4507 xexit (1);
4510 /* Provide an estimate of total stack required. */
4512 static bfd_boolean
4513 spu_elf_stack_analysis (struct bfd_link_info *info)
4515 struct spu_link_hash_table *htab;
4516 struct _sum_stack_param sum_stack_param;
4518 if (!discover_functions (info))
4519 return FALSE;
4521 if (!build_call_tree (info))
4522 return FALSE;
4524 htab = spu_hash_table (info);
4525 if (htab->params->stack_analysis)
4527 info->callbacks->info (_("Stack size for call graph root nodes.\n"));
4528 info->callbacks->minfo (_("\nStack size for functions. "
4529 "Annotations: '*' max stack, 't' tail call\n"));
4532 sum_stack_param.emit_stack_syms = htab->params->emit_stack_syms;
4533 sum_stack_param.overall_stack = 0;
4534 if (!for_each_node (sum_stack, info, &sum_stack_param, TRUE))
4535 return FALSE;
4537 if (htab->params->stack_analysis)
4538 info->callbacks->info (_("Maximum stack required is 0x%v\n"),
4539 (bfd_vma) sum_stack_param.overall_stack);
4540 return TRUE;
4543 /* Perform a final link. */
4545 static bfd_boolean
4546 spu_elf_final_link (bfd *output_bfd, struct bfd_link_info *info)
4548 struct spu_link_hash_table *htab = spu_hash_table (info);
4550 if (htab->params->auto_overlay)
4551 spu_elf_auto_overlay (info);
4553 if ((htab->params->stack_analysis
4554 || (htab->params->ovly_flavour == ovly_soft_icache
4555 && htab->params->lrlive_analysis))
4556 && !spu_elf_stack_analysis (info))
4557 info->callbacks->einfo ("%X%P: stack/lrlive analysis error: %E\n");
4559 if (!spu_elf_build_stubs (info))
4560 info->callbacks->einfo ("%F%P: can not build overlay stubs: %E\n");
4562 return bfd_elf_final_link (output_bfd, info);
4565 /* Called when not normally emitting relocs, ie. !info->relocatable
4566 and !info->emitrelocations. Returns a count of special relocs
4567 that need to be emitted. */
4569 static unsigned int
4570 spu_elf_count_relocs (struct bfd_link_info *info, asection *sec)
4572 Elf_Internal_Rela *relocs;
4573 unsigned int count = 0;
4575 relocs = _bfd_elf_link_read_relocs (sec->owner, sec, NULL, NULL,
4576 info->keep_memory);
4577 if (relocs != NULL)
4579 Elf_Internal_Rela *rel;
4580 Elf_Internal_Rela *relend = relocs + sec->reloc_count;
4582 for (rel = relocs; rel < relend; rel++)
4584 int r_type = ELF32_R_TYPE (rel->r_info);
4585 if (r_type == R_SPU_PPU32 || r_type == R_SPU_PPU64)
4586 ++count;
4589 if (elf_section_data (sec)->relocs != relocs)
4590 free (relocs);
4593 return count;
4596 /* Apply RELOCS to CONTENTS of INPUT_SECTION from INPUT_BFD. */
4598 static int
4599 spu_elf_relocate_section (bfd *output_bfd,
4600 struct bfd_link_info *info,
4601 bfd *input_bfd,
4602 asection *input_section,
4603 bfd_byte *contents,
4604 Elf_Internal_Rela *relocs,
4605 Elf_Internal_Sym *local_syms,
4606 asection **local_sections)
4608 Elf_Internal_Shdr *symtab_hdr;
4609 struct elf_link_hash_entry **sym_hashes;
4610 Elf_Internal_Rela *rel, *relend;
4611 struct spu_link_hash_table *htab;
4612 asection *ea;
4613 int ret = TRUE;
4614 bfd_boolean emit_these_relocs = FALSE;
4615 bfd_boolean is_ea_sym;
4616 bfd_boolean stubs;
4617 unsigned int iovl = 0;
4619 htab = spu_hash_table (info);
4620 stubs = (htab->stub_sec != NULL
4621 && maybe_needs_stubs (input_section));
4622 iovl = overlay_index (input_section);
4623 ea = bfd_get_section_by_name (output_bfd, "._ea");
4624 symtab_hdr = &elf_tdata (input_bfd)->symtab_hdr;
4625 sym_hashes = (struct elf_link_hash_entry **) (elf_sym_hashes (input_bfd));
4627 rel = relocs;
4628 relend = relocs + input_section->reloc_count;
4629 for (; rel < relend; rel++)
4631 int r_type;
4632 reloc_howto_type *howto;
4633 unsigned int r_symndx;
4634 Elf_Internal_Sym *sym;
4635 asection *sec;
4636 struct elf_link_hash_entry *h;
4637 const char *sym_name;
4638 bfd_vma relocation;
4639 bfd_vma addend;
4640 bfd_reloc_status_type r;
4641 bfd_boolean unresolved_reloc;
4642 bfd_boolean warned;
4643 bfd_boolean overlay_encoded;
4644 enum _stub_type stub_type;
4646 r_symndx = ELF32_R_SYM (rel->r_info);
4647 r_type = ELF32_R_TYPE (rel->r_info);
4648 howto = elf_howto_table + r_type;
4649 unresolved_reloc = FALSE;
4650 warned = FALSE;
4651 h = NULL;
4652 sym = NULL;
4653 sec = NULL;
4654 if (r_symndx < symtab_hdr->sh_info)
4656 sym = local_syms + r_symndx;
4657 sec = local_sections[r_symndx];
4658 sym_name = bfd_elf_sym_name (input_bfd, symtab_hdr, sym, sec);
4659 relocation = _bfd_elf_rela_local_sym (output_bfd, sym, &sec, rel);
4661 else
4663 if (sym_hashes == NULL)
4664 return FALSE;
4666 h = sym_hashes[r_symndx - symtab_hdr->sh_info];
4668 while (h->root.type == bfd_link_hash_indirect
4669 || h->root.type == bfd_link_hash_warning)
4670 h = (struct elf_link_hash_entry *) h->root.u.i.link;
4672 relocation = 0;
4673 if (h->root.type == bfd_link_hash_defined
4674 || h->root.type == bfd_link_hash_defweak)
4676 sec = h->root.u.def.section;
4677 if (sec == NULL
4678 || sec->output_section == NULL)
4679 /* Set a flag that will be cleared later if we find a
4680 relocation value for this symbol. output_section
4681 is typically NULL for symbols satisfied by a shared
4682 library. */
4683 unresolved_reloc = TRUE;
4684 else
4685 relocation = (h->root.u.def.value
4686 + sec->output_section->vma
4687 + sec->output_offset);
4689 else if (h->root.type == bfd_link_hash_undefweak)
4691 else if (info->unresolved_syms_in_objects == RM_IGNORE
4692 && ELF_ST_VISIBILITY (h->other) == STV_DEFAULT)
4694 else if (!info->relocatable
4695 && !(r_type == R_SPU_PPU32 || r_type == R_SPU_PPU64))
4697 bfd_boolean err;
4698 err = (info->unresolved_syms_in_objects == RM_GENERATE_ERROR
4699 || ELF_ST_VISIBILITY (h->other) != STV_DEFAULT);
4700 if (!info->callbacks->undefined_symbol (info,
4701 h->root.root.string,
4702 input_bfd,
4703 input_section,
4704 rel->r_offset, err))
4705 return FALSE;
4706 warned = TRUE;
4708 sym_name = h->root.root.string;
4711 if (sec != NULL && elf_discarded_section (sec))
4713 /* For relocs against symbols from removed linkonce sections,
4714 or sections discarded by a linker script, we just want the
4715 section contents zeroed. Avoid any special processing. */
4716 _bfd_clear_contents (howto, input_bfd, contents + rel->r_offset);
4717 rel->r_info = 0;
4718 rel->r_addend = 0;
4719 continue;
4722 if (info->relocatable)
4723 continue;
4725 is_ea_sym = (ea != NULL
4726 && sec != NULL
4727 && sec->output_section == ea);
4728 overlay_encoded = FALSE;
4730 /* If this symbol is in an overlay area, we may need to relocate
4731 to the overlay stub. */
4732 addend = rel->r_addend;
4733 if (stubs
4734 && !is_ea_sym
4735 && (stub_type = needs_ovl_stub (h, sym, sec, input_section, rel,
4736 contents, info)) != no_stub)
4738 unsigned int ovl = 0;
4739 struct got_entry *g, **head;
4741 if (stub_type != nonovl_stub)
4742 ovl = iovl;
4744 if (h != NULL)
4745 head = &h->got.glist;
4746 else
4747 head = elf_local_got_ents (input_bfd) + r_symndx;
4749 for (g = *head; g != NULL; g = g->next)
4750 if (htab->params->ovly_flavour == ovly_soft_icache
4751 ? g->br_addr == (rel->r_offset
4752 + input_section->output_offset
4753 + input_section->output_section->vma)
4754 : g->addend == addend && (g->ovl == ovl || g->ovl == 0))
4755 break;
4756 if (g == NULL)
4757 abort ();
4759 relocation = g->stub_addr;
4760 addend = 0;
4762 else
4764 /* For soft icache, encode the overlay index into addresses. */
4765 if (htab->params->ovly_flavour == ovly_soft_icache
4766 && !is_ea_sym)
4768 unsigned int ovl = overlay_index (sec);
4769 if (ovl != 0)
4771 unsigned int set_id = (ovl - 1) >> htab->num_lines_log2;
4772 relocation += set_id << 18;
4773 overlay_encoded = set_id != 0;
4778 if (unresolved_reloc)
4780 else if (r_type == R_SPU_PPU32 || r_type == R_SPU_PPU64)
4782 if (is_ea_sym)
4784 /* ._ea is a special section that isn't allocated in SPU
4785 memory, but rather occupies space in PPU memory as
4786 part of an embedded ELF image. If this reloc is
4787 against a symbol defined in ._ea, then transform the
4788 reloc into an equivalent one without a symbol
4789 relative to the start of the ELF image. */
4790 rel->r_addend += (relocation
4791 - ea->vma
4792 + elf_section_data (ea)->this_hdr.sh_offset);
4793 rel->r_info = ELF32_R_INFO (0, r_type);
4795 emit_these_relocs = TRUE;
4796 continue;
4798 else if (is_ea_sym)
4799 unresolved_reloc = TRUE;
4801 if (unresolved_reloc)
4803 (*_bfd_error_handler)
4804 (_("%B(%s+0x%lx): unresolvable %s relocation against symbol `%s'"),
4805 input_bfd,
4806 bfd_get_section_name (input_bfd, input_section),
4807 (long) rel->r_offset,
4808 howto->name,
4809 sym_name);
4810 ret = FALSE;
4813 r = _bfd_final_link_relocate (howto,
4814 input_bfd,
4815 input_section,
4816 contents,
4817 rel->r_offset, relocation, addend);
4819 if (r != bfd_reloc_ok)
4821 const char *msg = (const char *) 0;
4823 switch (r)
4825 case bfd_reloc_overflow:
4826 /* FIXME: We don't want to warn on most references
4827 within an overlay to itself, but this may silence a
4828 warning that should be reported. */
4829 if (overlay_encoded && sec == input_section)
4830 break;
4831 if (!((*info->callbacks->reloc_overflow)
4832 (info, (h ? &h->root : NULL), sym_name, howto->name,
4833 (bfd_vma) 0, input_bfd, input_section, rel->r_offset)))
4834 return FALSE;
4835 break;
4837 case bfd_reloc_undefined:
4838 if (!((*info->callbacks->undefined_symbol)
4839 (info, sym_name, input_bfd, input_section,
4840 rel->r_offset, TRUE)))
4841 return FALSE;
4842 break;
4844 case bfd_reloc_outofrange:
4845 msg = _("internal error: out of range error");
4846 goto common_error;
4848 case bfd_reloc_notsupported:
4849 msg = _("internal error: unsupported relocation error");
4850 goto common_error;
4852 case bfd_reloc_dangerous:
4853 msg = _("internal error: dangerous error");
4854 goto common_error;
4856 default:
4857 msg = _("internal error: unknown error");
4858 /* fall through */
4860 common_error:
4861 ret = FALSE;
4862 if (!((*info->callbacks->warning)
4863 (info, msg, sym_name, input_bfd, input_section,
4864 rel->r_offset)))
4865 return FALSE;
4866 break;
4871 if (ret
4872 && emit_these_relocs
4873 && !info->emitrelocations)
4875 Elf_Internal_Rela *wrel;
4876 Elf_Internal_Shdr *rel_hdr;
4878 wrel = rel = relocs;
4879 relend = relocs + input_section->reloc_count;
4880 for (; rel < relend; rel++)
4882 int r_type;
4884 r_type = ELF32_R_TYPE (rel->r_info);
4885 if (r_type == R_SPU_PPU32 || r_type == R_SPU_PPU64)
4886 *wrel++ = *rel;
4888 input_section->reloc_count = wrel - relocs;
4889 /* Backflips for _bfd_elf_link_output_relocs. */
4890 rel_hdr = &elf_section_data (input_section)->rel_hdr;
4891 rel_hdr->sh_size = input_section->reloc_count * rel_hdr->sh_entsize;
4892 ret = 2;
4895 return ret;
4898 /* Adjust _SPUEAR_ syms to point at their overlay stubs. */
4900 static bfd_boolean
4901 spu_elf_output_symbol_hook (struct bfd_link_info *info,
4902 const char *sym_name ATTRIBUTE_UNUSED,
4903 Elf_Internal_Sym *sym,
4904 asection *sym_sec ATTRIBUTE_UNUSED,
4905 struct elf_link_hash_entry *h)
4907 struct spu_link_hash_table *htab = spu_hash_table (info);
4909 if (!info->relocatable
4910 && htab->stub_sec != NULL
4911 && h != NULL
4912 && (h->root.type == bfd_link_hash_defined
4913 || h->root.type == bfd_link_hash_defweak)
4914 && h->def_regular
4915 && strncmp (h->root.root.string, "_SPUEAR_", 8) == 0)
4917 struct got_entry *g;
4919 for (g = h->got.glist; g != NULL; g = g->next)
4920 if (htab->params->ovly_flavour == ovly_soft_icache
4921 ? g->br_addr == g->stub_addr
4922 : g->addend == 0 && g->ovl == 0)
4924 sym->st_shndx = (_bfd_elf_section_from_bfd_section
4925 (htab->stub_sec[0]->output_section->owner,
4926 htab->stub_sec[0]->output_section));
4927 sym->st_value = g->stub_addr;
4928 break;
4932 return TRUE;
4935 static int spu_plugin = 0;
4937 void
4938 spu_elf_plugin (int val)
4940 spu_plugin = val;
4943 /* Set ELF header e_type for plugins. */
4945 static void
4946 spu_elf_post_process_headers (bfd *abfd,
4947 struct bfd_link_info *info ATTRIBUTE_UNUSED)
4949 if (spu_plugin)
4951 Elf_Internal_Ehdr *i_ehdrp = elf_elfheader (abfd);
4953 i_ehdrp->e_type = ET_DYN;
4957 /* We may add an extra PT_LOAD segment for .toe. We also need extra
4958 segments for overlays. */
4960 static int
4961 spu_elf_additional_program_headers (bfd *abfd, struct bfd_link_info *info)
4963 int extra = 0;
4964 asection *sec;
4966 if (info != NULL)
4968 struct spu_link_hash_table *htab = spu_hash_table (info);
4969 extra = htab->num_overlays;
4972 if (extra)
4973 ++extra;
4975 sec = bfd_get_section_by_name (abfd, ".toe");
4976 if (sec != NULL && (sec->flags & SEC_LOAD) != 0)
4977 ++extra;
4979 return extra;
4982 /* Remove .toe section from other PT_LOAD segments and put it in
4983 a segment of its own. Put overlays in separate segments too. */
4985 static bfd_boolean
4986 spu_elf_modify_segment_map (bfd *abfd, struct bfd_link_info *info)
4988 asection *toe, *s;
4989 struct elf_segment_map *m;
4990 unsigned int i;
4992 if (info == NULL)
4993 return TRUE;
4995 toe = bfd_get_section_by_name (abfd, ".toe");
4996 for (m = elf_tdata (abfd)->segment_map; m != NULL; m = m->next)
4997 if (m->p_type == PT_LOAD && m->count > 1)
4998 for (i = 0; i < m->count; i++)
4999 if ((s = m->sections[i]) == toe
5000 || spu_elf_section_data (s)->u.o.ovl_index != 0)
5002 struct elf_segment_map *m2;
5003 bfd_vma amt;
5005 if (i + 1 < m->count)
5007 amt = sizeof (struct elf_segment_map);
5008 amt += (m->count - (i + 2)) * sizeof (m->sections[0]);
5009 m2 = bfd_zalloc (abfd, amt);
5010 if (m2 == NULL)
5011 return FALSE;
5012 m2->count = m->count - (i + 1);
5013 memcpy (m2->sections, m->sections + i + 1,
5014 m2->count * sizeof (m->sections[0]));
5015 m2->p_type = PT_LOAD;
5016 m2->next = m->next;
5017 m->next = m2;
5019 m->count = 1;
5020 if (i != 0)
5022 m->count = i;
5023 amt = sizeof (struct elf_segment_map);
5024 m2 = bfd_zalloc (abfd, amt);
5025 if (m2 == NULL)
5026 return FALSE;
5027 m2->p_type = PT_LOAD;
5028 m2->count = 1;
5029 m2->sections[0] = s;
5030 m2->next = m->next;
5031 m->next = m2;
5033 break;
5036 return TRUE;
5039 /* Tweak the section type of .note.spu_name. */
5041 static bfd_boolean
5042 spu_elf_fake_sections (bfd *obfd ATTRIBUTE_UNUSED,
5043 Elf_Internal_Shdr *hdr,
5044 asection *sec)
5046 if (strcmp (sec->name, SPU_PTNOTE_SPUNAME) == 0)
5047 hdr->sh_type = SHT_NOTE;
5048 return TRUE;
5051 /* Tweak phdrs before writing them out. */
5053 static int
5054 spu_elf_modify_program_headers (bfd *abfd, struct bfd_link_info *info)
5056 const struct elf_backend_data *bed;
5057 struct elf_obj_tdata *tdata;
5058 Elf_Internal_Phdr *phdr, *last;
5059 struct spu_link_hash_table *htab;
5060 unsigned int count;
5061 unsigned int i;
5063 if (info == NULL)
5064 return TRUE;
5066 bed = get_elf_backend_data (abfd);
5067 tdata = elf_tdata (abfd);
5068 phdr = tdata->phdr;
5069 count = tdata->program_header_size / bed->s->sizeof_phdr;
5070 htab = spu_hash_table (info);
5071 if (htab->num_overlays != 0)
5073 struct elf_segment_map *m;
5074 unsigned int o;
5076 for (i = 0, m = elf_tdata (abfd)->segment_map; m; ++i, m = m->next)
5077 if (m->count != 0
5078 && (o = spu_elf_section_data (m->sections[0])->u.o.ovl_index) != 0)
5080 /* Mark this as an overlay header. */
5081 phdr[i].p_flags |= PF_OVERLAY;
5083 if (htab->ovtab != NULL && htab->ovtab->size != 0
5084 && htab->params->ovly_flavour != ovly_soft_icache)
5086 bfd_byte *p = htab->ovtab->contents;
5087 unsigned int off = o * 16 + 8;
5089 /* Write file_off into _ovly_table. */
5090 bfd_put_32 (htab->ovtab->owner, phdr[i].p_offset, p + off);
5093 /* Soft-icache has its file offset put in .ovl.init. */
5094 if (htab->init != NULL && htab->init->size != 0)
5096 bfd_vma val = elf_section_data (htab->ovl_sec[0])->this_hdr.sh_offset;
5098 bfd_put_32 (htab->init->owner, val, htab->init->contents + 4);
5102 /* Round up p_filesz and p_memsz of PT_LOAD segments to multiples
5103 of 16. This should always be possible when using the standard
5104 linker scripts, but don't create overlapping segments if
5105 someone is playing games with linker scripts. */
5106 last = NULL;
5107 for (i = count; i-- != 0; )
5108 if (phdr[i].p_type == PT_LOAD)
5110 unsigned adjust;
5112 adjust = -phdr[i].p_filesz & 15;
5113 if (adjust != 0
5114 && last != NULL
5115 && phdr[i].p_offset + phdr[i].p_filesz > last->p_offset - adjust)
5116 break;
5118 adjust = -phdr[i].p_memsz & 15;
5119 if (adjust != 0
5120 && last != NULL
5121 && phdr[i].p_filesz != 0
5122 && phdr[i].p_vaddr + phdr[i].p_memsz > last->p_vaddr - adjust
5123 && phdr[i].p_vaddr + phdr[i].p_memsz <= last->p_vaddr)
5124 break;
5126 if (phdr[i].p_filesz != 0)
5127 last = &phdr[i];
5130 if (i == (unsigned int) -1)
5131 for (i = count; i-- != 0; )
5132 if (phdr[i].p_type == PT_LOAD)
5134 unsigned adjust;
5136 adjust = -phdr[i].p_filesz & 15;
5137 phdr[i].p_filesz += adjust;
5139 adjust = -phdr[i].p_memsz & 15;
5140 phdr[i].p_memsz += adjust;
5143 return TRUE;
5146 #define TARGET_BIG_SYM bfd_elf32_spu_vec
5147 #define TARGET_BIG_NAME "elf32-spu"
5148 #define ELF_ARCH bfd_arch_spu
5149 #define ELF_MACHINE_CODE EM_SPU
5150 /* This matches the alignment need for DMA. */
5151 #define ELF_MAXPAGESIZE 0x80
5152 #define elf_backend_rela_normal 1
5153 #define elf_backend_can_gc_sections 1
5155 #define bfd_elf32_bfd_reloc_type_lookup spu_elf_reloc_type_lookup
5156 #define bfd_elf32_bfd_reloc_name_lookup spu_elf_reloc_name_lookup
5157 #define elf_info_to_howto spu_elf_info_to_howto
5158 #define elf_backend_count_relocs spu_elf_count_relocs
5159 #define elf_backend_relocate_section spu_elf_relocate_section
5160 #define elf_backend_symbol_processing spu_elf_backend_symbol_processing
5161 #define elf_backend_link_output_symbol_hook spu_elf_output_symbol_hook
5162 #define elf_backend_object_p spu_elf_object_p
5163 #define bfd_elf32_new_section_hook spu_elf_new_section_hook
5164 #define bfd_elf32_bfd_link_hash_table_create spu_elf_link_hash_table_create
5166 #define elf_backend_additional_program_headers spu_elf_additional_program_headers
5167 #define elf_backend_modify_segment_map spu_elf_modify_segment_map
5168 #define elf_backend_modify_program_headers spu_elf_modify_program_headers
5169 #define elf_backend_post_process_headers spu_elf_post_process_headers
5170 #define elf_backend_fake_sections spu_elf_fake_sections
5171 #define elf_backend_special_sections spu_elf_special_sections
5172 #define bfd_elf32_bfd_final_link spu_elf_final_link
5174 #include "elf32-target.h"