/** General API for VPR Other software tools should generally call just the functions defined here For advanced/power users, you can call functions defined elsewhere in VPR or modify the data structures directly at your discretion but be aware that doing so can break the correctness of VPR Author: Jason Luu June 21, 2012 */ #include #include #include #include #include "util.h" #include "vpr_types.h" #include "vpr_utils.h" #include "globals.h" #include "graphics.h" #include "read_netlist.h" #include "check_netlist.h" #include "print_netlist.h" #include "read_blif.h" #include "draw.h" #include "place_and_route.h" #include "pack.h" #include "SetupGrid.h" #include "stats.h" #include "path_delay.h" #include "OptionTokens.h" #include "ReadOptions.h" #include "read_xml_arch_file.h" #include "SetupVPR.h" #include "rr_graph.h" #include "pb_type_graph.h" #include "ReadOptions.h" #include "route_common.h" #include "timing_place_lookup.h" #include "cluster_legality.h" #include "route_export.h" #include "vpr_api.h" #include "read_sdc.h" #include "power.h" /* Xifan TANG: Add place_macro.h */ #include "place_macro.h" /* Local subroutines */ static void free_pb_type(t_pb_type *pb_type); static void free_complex_block_types(void); static void free_arch(t_arch* Arch); static void free_options(t_options *options); static void free_circuit(void); static boolean has_printhandler_pre_vpr = FALSE; /* For resync of clustered netlist to the post-route solution. This function adds local nets to cluster */ static void reload_intra_cluster_nets(t_pb *pb); static t_trace *alloc_and_load_final_routing_trace(); static t_trace *expand_routing_trace(t_trace *trace, int ivpack_net); static void print_complete_net_trace(t_trace* trace, const char *file_name); static void resync_post_route_netlist(); static void clay_logical_equivalence_handling(const t_arch *arch); static void clay_lut_input_rebalancing(int iblock, t_pb *pb); static void clay_reload_ble_locations(int iblock); static void resync_pb_graph_nodes_in_pb(t_pb_graph_node *pb_graph_node, t_pb *pb); /* Local subroutines end */ /* Display general VPR information */ void vpr_print_title(void) { vpr_printf(TIO_MESSAGE_INFO, "\n"); vpr_printf(TIO_MESSAGE_INFO, "VPR FPGA Placement and Routing.\n"); vpr_printf(TIO_MESSAGE_INFO, "Version: Version " VPR_VERSION "\n"); vpr_printf(TIO_MESSAGE_INFO, "Compiled: " __DATE__ ".\n"); vpr_printf(TIO_MESSAGE_INFO, "University of Toronto\n"); vpr_printf(TIO_MESSAGE_INFO, "vpr@eecg.utoronto.ca\n"); vpr_printf(TIO_MESSAGE_INFO, "Enhancements: mrFPGA, RRAM, SWSEG, FPGA-SPICE by Xifan TANG, EPFL-LSI, Univ. of Utah-LNIS \n"); vpr_printf(TIO_MESSAGE_INFO, "Enhancements: Synthesizable Verilog Support by Xifan TANG, EPFL-LSI, Univ. of Utah-LNIS\n"); vpr_printf(TIO_MESSAGE_INFO, "Enhancements: Bitstream Generator Support by Xifan TANG, EPFL-LSI, Univ. of Utah-LNIS\n"); vpr_printf(TIO_MESSAGE_INFO, "Enhancements: OPIN_TO_CB, CLB_PIN_REMAP by Xifan TANG, EPFL-LSI, Univ. of Utah-LNIS\n"); vpr_printf(TIO_MESSAGE_INFO, "xifan.tang@utah.edu\n"); vpr_printf(TIO_MESSAGE_INFO, "This is free open source code under MIT license.\n"); vpr_printf(TIO_MESSAGE_INFO, "\n"); } /* Display help screen */ void vpr_print_usage(void) { vpr_printf(TIO_MESSAGE_INFO, "Usage: vpr fpga_architecture.xml circuit_name [Options ...]\n"); vpr_printf(TIO_MESSAGE_INFO, "\n"); vpr_printf(TIO_MESSAGE_INFO, "General Options: [--nodisp] [--auto ] [--pack]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--place] [--route] [--timing_analyze_only_with_net_delay ]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--fast] [--full_stats] [--timing_analysis on | off] [--outfile_prefix ]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--blif_file ][--net_file ][--place_file ]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--route_file ][--sdc_file ][--echo_file on | off]\n"); vpr_printf(TIO_MESSAGE_INFO, "\n"); vpr_printf(TIO_MESSAGE_INFO, "Packer Options:\n"); /* vpr_printf(TIO_MESSAGE_INFO, "\t[-global_clocks on|off]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[-hill_climbing on|off]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[-sweep_hanging_nets_and_inputs on|off]\n"); */ vpr_printf(TIO_MESSAGE_INFO, "\t[--timing_driven_clustering on|off]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--cluster_seed_type timing|max_inputs] [--alpha_clustering ] [--beta_clustering ]\n"); /* vpr_printf(TIO_MESSAGE_INFO, "\t[-recompute_timing_after ] [-cluster_block_delay ]\n"); */ vpr_printf(TIO_MESSAGE_INFO, "\t[--allow_unrelated_clustering on|off]\n"); /* vpr_printf(TIO_MESSAGE_INFO, "\t[-allow_early_exit on|off]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[-intra_cluster_net_delay ] \n"); vpr_printf(TIO_MESSAGE_INFO, "\t[-inter_cluster_net_delay ] \n"); */ vpr_printf(TIO_MESSAGE_INFO, "\t[--connection_driven_clustering on|off] \n"); vpr_printf(TIO_MESSAGE_INFO, "\n"); vpr_printf(TIO_MESSAGE_INFO, "Placer Options:\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--place_algorithm bounding_box | net_timing_driven | path_timing_driven]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--init_t ] [--exit_t ]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--alpha_t ] [--inner_num ] [--seed ]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--place_cost_exp ]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--place_chan_width ] \n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--fix_pins random | ]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--enable_timing_computations on | off]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--block_dist ]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--place_clb_pin_remap]\n"); vpr_printf(TIO_MESSAGE_INFO, "\n"); vpr_printf(TIO_MESSAGE_INFO, "Placement Options Valid Only for Timing-Driven Placement:\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--timing_tradeoff ]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--recompute_crit_iter ]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--inner_loop_recompute_divider ]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--td_place_exp_first ]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--td_place_exp_last ]\n"); vpr_printf(TIO_MESSAGE_INFO, "\n"); vpr_printf(TIO_MESSAGE_INFO, "Router Options: [-max_router_iterations ] [-bb_factor ]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--initial_pres_fac ] [--pres_fac_mult ]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--acc_fac ] [--first_iter_pres_fac ]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--bend_cost ] [--route_type global | detailed]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--verify_binary_search] [--route_chan_width ]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--router_algorithm breadth_first | timing_driven]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--base_cost_type intrinsic_delay | delay_normalized | demand_only]\n"); vpr_printf(TIO_MESSAGE_INFO, "\n"); vpr_printf(TIO_MESSAGE_INFO, "Routing options valid only for timing-driven routing:\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--astar_fac ] [--max_criticality ]\n"); vpr_printf(TIO_MESSAGE_INFO, "\t[--criticality_exp ]\n"); vpr_printf(TIO_MESSAGE_INFO, "\n"); vpr_printf(TIO_MESSAGE_INFO, "Power Options:\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--power\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--power_output_file \n"); vpr_printf(TIO_MESSAGE_INFO, "\t--activity_file \n"); vpr_printf(TIO_MESSAGE_INFO, "\t--tech_properties \n"); vpr_printf(TIO_MESSAGE_INFO, "\n"); /* Xifan TANG: FPGA-SPICE Support*/ vpr_printf(TIO_MESSAGE_INFO, "FPGA-X2P (from XML to Product/Prototype) tool suite Options:\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_x2p_rename_illegal_port\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_x2p_signal_density_weight \n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_x2p_sim_window_size \n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_x2p_compact_routing_hierarchy\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_x2p_output_sb_xml \n"); vpr_printf(TIO_MESSAGE_INFO, "SPICE Support Options:\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_spice\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_spice_dir \n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_spice_print_top_testbench\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_spice_print_lut_testbench\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_spice_print_hardlogic_testbench\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_spice_print_io_testbench\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_spice_print_pb_mux_testbench\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_spice_print_cb_mux_testbench\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_spice_print_sb_mux_testbench\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_spice_print_cb_testbench\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_spice_print_sb_testbench\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_spice_print_grid_testbench\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_spice_leakage_only\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_spice_parasitic_net_estimation \n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_spice_testbench_load_extraction \n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_spice_sim_mt_num \n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_spice_simulator_path \n"); /* Xifan TANG: Synthesizable Verilog Dump*/ vpr_printf(TIO_MESSAGE_INFO, "Synthesizable Verilog Generator Options:\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_verilog\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_verilog_dir \n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_verilog_explicit_mapping\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_verilog_print_top_testbench\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_verilog_print_autocheck_top_testbench \n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_verilog_print_input_blif_testbench\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_verilog_print_formal_verification_top_netlist\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_verilog_include_timing\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_verilog_include_signal_init\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_verilog_include_icarus_simulator\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_verilog_print_modelsim_autodeck \n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_verilog_print_user_defined_template\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_verilog_print_report_timing_tcl\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_verilog_report_timing_rpt_path \n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_verilog_print_sdc_pnr\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_verilog_print_sdc_analysis\n"); /* Xifan Tang: Bitstream generator */ vpr_printf(TIO_MESSAGE_INFO, "Bitstream Generator Options:\n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_bitstream_generator \n"); vpr_printf(TIO_MESSAGE_INFO, "\t--fpga_bitstream_output_file \n"); } void vpr_init_file_handler() { char* pszLogFileName = "vpr_stdout.log"; unsigned char enableTimeStamps = 1; unsigned long maxWarningCount = 100000; unsigned long maxErrorCount = 1000; if (PrintHandlerExists() == 1) { has_printhandler_pre_vpr = TRUE; } else { has_printhandler_pre_vpr = FALSE; } if (has_printhandler_pre_vpr == FALSE) { PrintHandlerNew(pszLogFileName); PrintHandlerInit(enableTimeStamps, maxWarningCount, maxErrorCount); } return; } /* Initialize VPR 1. Read Options 2. Read Arch 3. Read Circuit 4. Sanity check all three */ void vpr_init(INP int argc, INP char **argv, OUTP t_options *options, OUTP t_vpr_setup *vpr_setup, OUTP t_arch *arch) { /* Initialize file handler */ vpr_init_file_handler(); /* Print title message */ vpr_print_title(); /* Print usage message if no args */ if (argc < 3) { vpr_print_usage(); exit(1); } memset(options, 0, sizeof(t_options)); memset(vpr_setup, 0, sizeof(t_vpr_setup)); memset(arch, 0, sizeof(t_arch)); /* Read in user options */ ReadOptions(argc, argv, options); /* Timing option priorities */ vpr_setup->TimingEnabled = IsTimingEnabled(options); /* Determine whether echo is on or off */ setEchoEnabled(IsEchoEnabled(options)); SetPostSynthesisOption(IsPostSynthesisEnabled(options)); vpr_setup->constant_net_delay = options->constant_net_delay; /* Read in arch and circuit */ SetupVPR(options, vpr_setup->TimingEnabled, TRUE, &vpr_setup->FileNameOpts, arch, &vpr_setup->Operation, &vpr_setup->user_models, &vpr_setup->library_models, &vpr_setup->PackerOpts, &vpr_setup->PlacerOpts, &vpr_setup->AnnealSched, &vpr_setup->RouterOpts, &vpr_setup->RoutingArch, &vpr_setup->Segments, &vpr_setup->Timing, &vpr_setup->ShowGraphics, &vpr_setup->GraphPause, &vpr_setup->PowerOpts, /*Xifan TANG: Switch Segment Pattern Support*/ &vpr_setup->swseg_patterns, &vpr_setup->FPGA_SPICE_Opts); /* Check inputs are reasonable */ CheckOptions(*options, vpr_setup->TimingEnabled); CheckArch(*arch, vpr_setup->TimingEnabled); /* Verify settings don't conflict or otherwise not make sense */ CheckSetup(vpr_setup->Operation, vpr_setup->PlacerOpts, vpr_setup->AnnealSched, vpr_setup->RouterOpts, vpr_setup->RoutingArch, vpr_setup->Segments, vpr_setup->Timing, arch->Chans); /* flush any messages to user still in stdout that hasn't gotten displayed */ fflush(stdout); /* Read blif file and sweep unused components */ read_and_process_blif(vpr_setup->PackerOpts.blif_file_name, vpr_setup->PackerOpts.sweep_hanging_nets_and_inputs, vpr_setup->user_models, vpr_setup->library_models, /* Xifan TANG: we need activity in spice modeling */ (boolean)(vpr_setup->PowerOpts.do_power | vpr_setup->FPGA_SPICE_Opts.read_act_file), vpr_setup->FileNameOpts.ActFile); fflush(stdout); ShowSetup(*options, *vpr_setup); } /* * Sets globals: nx, ny * Allocs globals: chan_width_x, chan_width_y, grid * Depends on num_clbs, pins_per_clb */ void vpr_init_pre_place_and_route(INP t_vpr_setup vpr_setup, INP t_arch Arch) { int *num_instances_type, *num_blocks_type; int i; int current, high, low; boolean fit; /* Xifan TANG: consider marco length in fitting */ int imacro, max_len_chain_blocks; int num_pl_macros; t_pl_macro* pl_macros; /* Read in netlist file for placement and routing */ if (vpr_setup.FileNameOpts.NetFile) { read_netlist(vpr_setup.FileNameOpts.NetFile, &Arch, &num_blocks, &block, &num_nets, &clb_net); /* This is done so that all blocks have subblocks and can be treated the same */ check_netlist(); } /* Output the current settings to console. */ printClusteredNetlistStats(); if (vpr_setup.Operation == TIMING_ANALYSIS_ONLY) { do_constant_net_delay_timing_analysis(vpr_setup.Timing, vpr_setup.constant_net_delay); } else { current = nint((float)sqrt((float)num_blocks)); /* current is the value of the smaller side of the FPGA */ low = 1; high = -1; num_instances_type = (int*) my_calloc(num_types, sizeof(int)); num_blocks_type = (int*) my_calloc(num_types, sizeof(int)); for (i = 0; i < num_blocks; i++) { num_blocks_type[block[i].type->index]++; } if (Arch.clb_grid.IsAuto) { /* Auto-size FPGA, perform a binary search */ while (high == -1 || low < high) { /* Generate grid */ if (Arch.clb_grid.Aspect >= 1.0) { ny = current; nx = nint(current * Arch.clb_grid.Aspect); } else { nx = current; ny = nint(current / Arch.clb_grid.Aspect); } #if DEBUG vpr_printf(TIO_MESSAGE_INFO, "Auto-sizing FPGA at x = %d y = %d\n", nx, ny); #endif alloc_and_load_grid(num_instances_type); freeGrid(); /* Xifan TANG: We need consider the length of carry-chain CLBs into account! */ num_pl_macros = alloc_and_load_placement_macros(Arch.Directs, Arch.num_directs, &pl_macros); /* find length of longest carry-chain logic blocks */ max_len_chain_blocks = max_len_pl_macros(num_pl_macros, pl_macros); /* Free all the allocated structs */ free_placement_macros_structs(); for (imacro = 0; imacro < num_pl_macros; imacro ++) { free(pl_macros[imacro].members); } free(pl_macros); /* Test if netlist fits in grid */ fit = TRUE; for (i = 0; i < num_types; i++) { if (num_blocks_type[i] > num_instances_type[i]) { fit = FALSE; break; } } /* If the length of macros is longer than ny - 2, fitting should fail. * Note: carry-chain logic blocks are placed only vertically in FPGA. */ if ((TRUE == fit)&&(max_len_chain_blocks > (ny))) { fit = FALSE; vpr_printf(TIO_MESSAGE_INFO, "Carry-chain logic blocks length (%d) is larger than y (%d)!\n", max_len_chain_blocks, ny); } /* get next value */ if (!fit) { /* increase size of max */ if (high == -1) { current = current * 2; if (current > MAX_SHORT) { vpr_printf(TIO_MESSAGE_ERROR, "FPGA required is too large for current architecture settings.\n"); exit(1); } } else { if (low == current) current++; low = current; current = low + ((high - low) / 2); } } else { high = current; current = low + ((high - low) / 2); } } /* Generate grid */ if (Arch.clb_grid.Aspect >= 1.0) { ny = current; nx = nint(current * Arch.clb_grid.Aspect); } else { nx = current; ny = nint(current / Arch.clb_grid.Aspect); } alloc_and_load_grid(num_instances_type); vpr_printf(TIO_MESSAGE_INFO, "FPGA auto-sized to x = %d y = %d\n", nx, ny); } else { nx = Arch.clb_grid.W; ny = Arch.clb_grid.H; alloc_and_load_grid(num_instances_type); } vpr_printf(TIO_MESSAGE_INFO, "The circuit will be mapped into a %d x %d array of clbs.\n", nx, ny); /* Xifan TANG: We need consider the length of carry-chain CLBs into account! */ num_pl_macros = alloc_and_load_placement_macros(Arch.Directs, Arch.num_directs, &pl_macros); /* find length of longest carry-chain logic blocks */ max_len_chain_blocks = max_len_pl_macros(num_pl_macros, pl_macros); /* Free all the allocated structs */ free_placement_macros_structs(); for (imacro = 0; imacro < num_pl_macros; imacro ++) { free(pl_macros[imacro].members); } free(pl_macros); /* Test if netlist fits in grid */ fit = TRUE; for (i = 0; i < num_types; i++) { if (num_blocks_type[i] > num_instances_type[i]) { fit = FALSE; break; } } /* If the length of macros is longer than ny - 2, fitting should fail. * Note: carry-chain logic blocks are placed only vertically in FPGA. */ if ((TRUE == fit)&&(max_len_chain_blocks > (ny))) { fit = FALSE; vpr_printf(TIO_MESSAGE_INFO, "Carry-chain logic blocks length (%d) is larger than y (%d) !\n", max_len_chain_blocks, ny); } if (!fit) { vpr_printf(TIO_MESSAGE_ERROR, "Not enough physical locations for type %s, number of blocks is %d but number of locations is %d.\n", type_descriptors[i].name, num_blocks_type[i], num_instances_type[i]); exit(1); } vpr_printf(TIO_MESSAGE_INFO, "\n"); vpr_printf(TIO_MESSAGE_INFO, "Resource usage...\n"); for (i = 0; i < num_types; i++) { vpr_printf(TIO_MESSAGE_INFO, "\tNetlist %d\tblocks of type: %s\n", num_blocks_type[i], type_descriptors[i].name); vpr_printf(TIO_MESSAGE_INFO, "\tArchitecture %d\tblocks of type: %s\n", num_instances_type[i], type_descriptors[i].name); } vpr_printf(TIO_MESSAGE_INFO, "\n"); chan_width_x = (int *) my_malloc((ny + 1) * sizeof(int)); chan_width_y = (int *) my_malloc((nx + 1) * sizeof(int)); free(num_blocks_type); free(num_instances_type); } } void vpr_pack(INP t_vpr_setup vpr_setup, INP t_arch arch) { clock_t begin, end; float inter_cluster_delay = UNDEFINED, Tdel_opin_switch, Tdel_wire_switch, Tdel_wtoi_switch, R_opin_switch, R_wire_switch, R_wtoi_switch, Cout_opin_switch, Cout_wire_switch, Cout_wtoi_switch, opin_switch_del, wire_switch_del, wtoi_switch_del, Rmetal, Cmetal, first_wire_seg_delay, second_wire_seg_delay; begin = clock(); vpr_printf(TIO_MESSAGE_INFO, "Initialize packing.\n"); /* If needed, estimate inter-cluster delay. Assume the average routing hop goes out of a block through an opin switch to a length-4 wire, then through a wire switch to another length-4 wire, then through a wire-to-ipin-switch into another block. */ if (vpr_setup.PackerOpts.timing_driven && vpr_setup.PackerOpts.auto_compute_inter_cluster_net_delay) { opin_switch_del = get_switch_info(arch.Segments[0].opin_switch, Tdel_opin_switch, R_opin_switch, Cout_opin_switch); wire_switch_del = get_switch_info(arch.Segments[0].wire_switch, Tdel_wire_switch, R_wire_switch, Cout_wire_switch); wtoi_switch_del = get_switch_info( vpr_setup.RoutingArch.wire_to_ipin_switch, Tdel_wtoi_switch, R_wtoi_switch, Cout_wtoi_switch); /* wire-to-ipin switch */ Rmetal = arch.Segments[0].Rmetal; Cmetal = arch.Segments[0].Cmetal; /* The delay of a wire with its driving switch is the switch delay plus the product of the equivalent resistance and capacitance experienced by the wire. */ #define WIRE_SEGMENT_LENGTH 4 first_wire_seg_delay = opin_switch_del + (R_opin_switch + Rmetal * WIRE_SEGMENT_LENGTH / 2) * (Cout_opin_switch + Cmetal * WIRE_SEGMENT_LENGTH); second_wire_seg_delay = wire_switch_del + (R_wire_switch + Rmetal * WIRE_SEGMENT_LENGTH / 2) * (Cout_wire_switch + Cmetal * WIRE_SEGMENT_LENGTH); inter_cluster_delay = 4 * (first_wire_seg_delay + second_wire_seg_delay + wtoi_switch_del); /* multiply by 4 to get a more conservative estimate */ } try_pack(&vpr_setup.PackerOpts, &arch, vpr_setup.user_models, vpr_setup.library_models, vpr_setup.Timing, inter_cluster_delay); end = clock(); #ifdef CLOCKS_PER_SEC vpr_printf(TIO_MESSAGE_INFO, "Packing took %g seconds.\n", (float) (end - begin) / CLOCKS_PER_SEC); vpr_printf(TIO_MESSAGE_INFO, "Packing completed.\n"); #else vpr_printf(TIO_MESSAGE_INFO, "Packing took %g seconds.\n", (float)(end - begin) / CLK_PER_SEC); #endif /* Xifan TANG: print the run time of packing placement */ vpr_printf(TIO_MESSAGE_INFO, "Packing routing took %g seconds.\n", pack_route_time); fflush(stdout); } void vpr_place_and_route(INP t_vpr_setup vpr_setup, INP t_arch arch) { /* Startup X graphics */ set_graphics_state(vpr_setup.ShowGraphics, vpr_setup.GraphPause, vpr_setup.RouterOpts.route_type); if (vpr_setup.ShowGraphics) { init_graphics("VPR: Versatile Place and Route for FPGAs", WHITE); alloc_draw_structs(); } /* Do placement and routing */ place_and_route(vpr_setup.Operation, vpr_setup.PlacerOpts, vpr_setup.FileNameOpts.PlaceFile, vpr_setup.FileNameOpts.NetFile, vpr_setup.FileNameOpts.ArchFile, vpr_setup.FileNameOpts.RouteFile, vpr_setup.AnnealSched, vpr_setup.RouterOpts, vpr_setup.RoutingArch, vpr_setup.Segments, vpr_setup.Timing, arch.Chans, arch.models, arch.Directs, arch.num_directs, arch.sram_inf.area, /*Xifan TANG: Switch Segment Pattern Support*/ vpr_setup.swseg_patterns); fflush(stdout); /* Close down X Display */ /* TODO: DANGEROUS way of coding, clean up */ if (vpr_setup.ShowGraphics) close_graphics(); free_draw_structs(); } /* Free architecture data structures */ void free_arch(t_arch* Arch) { int i; t_model *model, *prev; t_model_ports *port, *prev_port; struct s_linked_vptr *vptr, *vptr_prev; freeGrid(); free(chan_width_x); chan_width_x = NULL; free(chan_width_y); chan_width_y = NULL; for (i = 0; i < Arch->num_switches; i++) { if (Arch->Switches->name != NULL) { free(Arch->Switches[i].name); } } free(Arch->Switches); free(switch_inf); for (i = 0; i < Arch->num_segments; i++) { if (Arch->Segments->cb != NULL) { free(Arch->Segments[i].cb); } if (Arch->Segments->sb != NULL) { free(Arch->Segments[i].sb); } } free(Arch->Segments); model = Arch->models; while (model) { port = model->inputs; while (port) { prev_port = port; port = port->next; free(prev_port->name); free(prev_port); } port = model->outputs; while (port) { prev_port = port; port = port->next; free(prev_port->name); free(prev_port); } vptr = model->pb_types; while (vptr) { vptr_prev = vptr; vptr = vptr->next; free(vptr_prev); } prev = model; model = model->next; if (prev->instances) free(prev->instances); free(prev->name); free(prev); } for (i = 0; i < 4; i++) { vptr = Arch->model_library[i].pb_types; while (vptr) { vptr_prev = vptr; vptr = vptr->next; free(vptr_prev); } } for (i = 0; i < Arch->num_directs; i++) { free(Arch->Directs[i].name); free(Arch->Directs[i].from_pin); free(Arch->Directs[i].to_pin); } free(Arch->Directs); free(Arch->model_library[0].name); free(Arch->model_library[0].outputs->name); free(Arch->model_library[0].outputs); free(Arch->model_library[1].inputs->name); free(Arch->model_library[1].inputs); free(Arch->model_library[1].name); free(Arch->model_library[2].name); free(Arch->model_library[2].inputs[0].name); free(Arch->model_library[2].inputs[1].name); free(Arch->model_library[2].inputs); free(Arch->model_library[2].outputs->name); free(Arch->model_library[2].outputs); free(Arch->model_library[3].name); free(Arch->model_library[3].inputs->name); free(Arch->model_library[3].inputs); free(Arch->model_library[3].outputs->name); free(Arch->model_library[3].outputs); free(Arch->model_library); if (Arch->clocks) { free(Arch->clocks->clock_inf); } free_complex_block_types(); free_chunk_memory_trace(); } void free_options(t_options *options) { free(options->ArchFile); free(options->CircuitName); if (options->ActFile) free(options->ActFile); if (options->BlifFile) free(options->BlifFile); if (options->NetFile) free(options->NetFile); if (options->PlaceFile) free(options->PlaceFile); if (options->PowerFile) free(options->PowerFile); if (options->CmosTechFile) free(options->CmosTechFile); if (options->RouteFile) free(options->RouteFile); if (options->out_file_prefix) free(options->out_file_prefix); if (options->PinFile) free(options->PinFile); } static void free_complex_block_types(void) { int i, j, k, m; free_all_pb_graph_nodes(); for (i = 0; i < num_types; i++) { if (&type_descriptors[i] == EMPTY_TYPE) { continue; } free(type_descriptors[i].name); for (j = 0; j < type_descriptors[i].height; j++) { for (k = 0; k < 4; k++) { for (m = 0; m < type_descriptors[i].num_pin_loc_assignments[j][k]; m++) { if (type_descriptors[i].pin_loc_assignments[j][k][m]) free(type_descriptors[i].pin_loc_assignments[j][k][m]); } free(type_descriptors[i].pinloc[j][k]); free(type_descriptors[i].pin_loc_assignments[j][k]); } free(type_descriptors[i].pinloc[j]); free(type_descriptors[i].pin_loc_assignments[j]); free(type_descriptors[i].num_pin_loc_assignments[j]); } for (j = 0; j < type_descriptors[i].num_class; j++) { free(type_descriptors[i].class_inf[j].pinlist); } free(type_descriptors[i].pinloc); free(type_descriptors[i].pin_loc_assignments); free(type_descriptors[i].num_pin_loc_assignments); free(type_descriptors[i].pin_height); free(type_descriptors[i].class_inf); free(type_descriptors[i].is_global_pin); free(type_descriptors[i].pin_class); free(type_descriptors[i].grid_loc_def); free(type_descriptors[i].is_Fc_frac); free(type_descriptors[i].is_Fc_full_flex); free(type_descriptors[i].Fc); free_pb_type(type_descriptors[i].pb_type); free(type_descriptors[i].pb_type); } free(type_descriptors); } static void free_pb_type(t_pb_type *pb_type) { int i, j, k, m; free(pb_type->name); if (pb_type->blif_model) free(pb_type->blif_model); for (i = 0; i < pb_type->num_modes; i++) { for (j = 0; j < pb_type->modes[i].num_pb_type_children; j++) { free_pb_type(&pb_type->modes[i].pb_type_children[j]); } free(pb_type->modes[i].pb_type_children); free(pb_type->modes[i].name); for (j = 0; j < pb_type->modes[i].num_interconnect; j++) { free(pb_type->modes[i].interconnect[j].input_string); free(pb_type->modes[i].interconnect[j].output_string); free(pb_type->modes[i].interconnect[j].name); for (k = 0; k < pb_type->modes[i].interconnect[j].num_annotations; k++) { if (pb_type->modes[i].interconnect[j].annotations[k].clock) free( pb_type->modes[i].interconnect[j].annotations[k].clock); if (pb_type->modes[i].interconnect[j].annotations[k].input_pins) { free( pb_type->modes[i].interconnect[j].annotations[k].input_pins); } if (pb_type->modes[i].interconnect[j].annotations[k].output_pins) { free( pb_type->modes[i].interconnect[j].annotations[k].output_pins); } for (m = 0; m < pb_type->modes[i].interconnect[j].annotations[k].num_value_prop_pairs; m++) { free( pb_type->modes[i].interconnect[j].annotations[k].value[m]); } free(pb_type->modes[i].interconnect[j].annotations[k].prop); free(pb_type->modes[i].interconnect[j].annotations[k].value); } free(pb_type->modes[i].interconnect[j].annotations); if (pb_type->modes[i].interconnect[j].interconnect_power) free(pb_type->modes[i].interconnect[j].interconnect_power); } if (pb_type->modes[i].interconnect) free(pb_type->modes[i].interconnect); if (pb_type->modes[i].mode_power) free(pb_type->modes[i].mode_power); } if (pb_type->modes) free(pb_type->modes); for (i = 0; i < pb_type->num_annotations; i++) { for (j = 0; j < pb_type->annotations[i].num_value_prop_pairs; j++) { free(pb_type->annotations[i].value[j]); } free(pb_type->annotations[i].value); free(pb_type->annotations[i].prop); if (pb_type->annotations[i].input_pins) { free(pb_type->annotations[i].input_pins); } if (pb_type->annotations[i].output_pins) { free(pb_type->annotations[i].output_pins); } if (pb_type->annotations[i].clock) { free(pb_type->annotations[i].clock); } } if (pb_type->num_annotations > 0) { free(pb_type->annotations); } if (pb_type->pb_type_power) { free(pb_type->pb_type_power); } for (i = 0; i < pb_type->num_ports; i++) { free(pb_type->ports[i].name); if (pb_type->ports[i].port_class) { free(pb_type->ports[i].port_class); } if (pb_type->ports[i].port_power) { free(pb_type->ports[i].port_power); } } free(pb_type->ports); } void free_circuit() { int i; struct s_linked_vptr *p_io_removed; /* Free netlist reference tables for nets */ free(clb_to_vpack_net_mapping); free(vpack_to_clb_net_mapping); clb_to_vpack_net_mapping = NULL; vpack_to_clb_net_mapping = NULL; /* Free logical blocks and nets */ if (logical_block != NULL) { free_logical_blocks(); free_logical_nets(); } if (clb_net != NULL) { for (i = 0; i < num_nets; i++) { free(clb_net[i].name); free(clb_net[i].node_block); free(clb_net[i].node_block_pin); free(clb_net[i].node_block_port); } } free(clb_net); clb_net = NULL; if (block != NULL) { for (i = 0; i < num_blocks; i++) { if (block[i].pb != NULL) { free_cb(block[i].pb); free(block[i].pb); } free(block[i].nets); free(block[i].name); } } free(block); block = NULL; free(blif_circuit_name); free(default_output_name); blif_circuit_name = NULL; p_io_removed = circuit_p_io_removed; while (p_io_removed != NULL) { circuit_p_io_removed = p_io_removed->next; free(p_io_removed->data_vptr); free(p_io_removed); p_io_removed = circuit_p_io_removed; } } void vpr_free_vpr_data_structures(INOUTP t_arch Arch, INOUTP t_options options, INOUTP t_vpr_setup vpr_setup) { if (vpr_setup.Timing.SDCFile != NULL) { free(vpr_setup.Timing.SDCFile); vpr_setup.Timing.SDCFile = NULL; } free_options(&options); free_circuit(); free_arch(&Arch); free_echo_file_info(); free_output_file_names(); free_timing_stats(); free_sdc_related_structs(); } void vpr_free_all(INOUTP t_arch Arch, INOUTP t_options options, INOUTP t_vpr_setup vpr_setup) { free_rr_graph(); if (vpr_setup.RouterOpts.doRouting) { free_route_structs(); } free_trace_structs(); vpr_free_vpr_data_structures(Arch, options, vpr_setup); if (has_printhandler_pre_vpr == FALSE) { PrintHandlerDelete(); } } /**************************************************************************************************** * Advanced functions * Used when you need fine-grained control over VPR that the main VPR operations do not enable ****************************************************************************************************/ /* Read in user options */ void vpr_read_options(INP int argc, INP char **argv, OUTP t_options * options) { ReadOptions(argc, argv, options); } /* Read in arch and circuit */ void vpr_setup_vpr(INP t_options *Options, INP boolean TimingEnabled, INP boolean readArchFile, OUTP struct s_file_name_opts *FileNameOpts, INOUTP t_arch * Arch, OUTP enum e_operation *Operation, OUTP t_model ** user_models, OUTP t_model ** library_models, OUTP struct s_packer_opts *PackerOpts, OUTP struct s_placer_opts *PlacerOpts, OUTP struct s_annealing_sched *AnnealSched, OUTP struct s_router_opts *RouterOpts, OUTP struct s_det_routing_arch *RoutingArch, OUTP t_segment_inf ** Segments, OUTP t_timing_inf * Timing, OUTP boolean * ShowGraphics, OUTP int *GraphPause, t_power_opts * PowerOpts, /*Xifan TANG: Switch Segment Pattern Support*/ OUTP t_swseg_pattern_inf** swseg_patterns, /* Xifan TANG: FPGA-SPICE Tool Suites Support*/ OUTP t_fpga_spice_opts* FPGA_SPICE_Opts) { SetupVPR(Options, TimingEnabled, readArchFile, FileNameOpts, Arch, Operation, user_models, library_models, PackerOpts, PlacerOpts, AnnealSched, RouterOpts, RoutingArch, Segments, Timing, ShowGraphics, GraphPause, PowerOpts, swseg_patterns, FPGA_SPICE_Opts); } /* Check inputs are reasonable */ void vpr_check_options(INP t_options Options, INP boolean TimingEnabled) { CheckOptions(Options, TimingEnabled); } void vpr_check_arch(INP t_arch Arch, INP boolean TimingEnabled) { CheckArch(Arch, TimingEnabled); } /* Verify settings don't conflict or otherwise not make sense */ void vpr_check_setup(INP enum e_operation Operation, INP struct s_placer_opts PlacerOpts, INP struct s_annealing_sched AnnealSched, INP struct s_router_opts RouterOpts, INP struct s_det_routing_arch RoutingArch, INP t_segment_inf * Segments, INP t_timing_inf Timing, INP t_chan_width_dist Chans) { CheckSetup(Operation, PlacerOpts, AnnealSched, RouterOpts, RoutingArch, Segments, Timing, Chans); } /* Read blif file and sweep unused components */ void vpr_read_and_process_blif(INP char *blif_file, INP boolean sweep_hanging_nets_and_inputs, INP t_model *user_models, INP t_model *library_models, boolean read_activity_file, char * activity_file) { read_and_process_blif(blif_file, sweep_hanging_nets_and_inputs, user_models, library_models, read_activity_file, activity_file); } /* Show current setup */ void vpr_show_setup(INP t_options options, INP t_vpr_setup vpr_setup) { ShowSetup(options, vpr_setup); } /* Output file names management */ void vpr_alloc_and_load_output_file_names(const char* default_name) { alloc_and_load_output_file_names(default_name); } void vpr_set_output_file_name(enum e_output_files ename, const char *name, const char* default_name) { setOutputFileName(ename, name, default_name); } char *vpr_get_output_file_name(enum e_output_files ename) { return getOutputFileName(ename); } /* logical equivalence scrambles the packed netlist indices with the actual indices, need to resync then re-output clustered netlist, this code assumes I'm dealing with a TI CLAY v1 architecture */ /* Returns a trace array [0..num_logical_nets-1] with the final routing of the circuit from the logical_block netlist, index of the trace array corresponds to the index of a vpack_net */ t_trace* vpr_resync_post_route_netlist_to_TI_CLAY_v1_architecture( INP const t_arch *arch) { t_trace *trace; /* Map post-routed traces to clb_nets and block */ resync_post_route_netlist(); /* Resolve logically equivalent inputs */ clay_logical_equivalence_handling(arch); /* Finalize traceback */ trace = alloc_and_load_final_routing_trace(); if (getEchoEnabled() && isEchoFileEnabled(E_ECHO_COMPLETE_NET_TRACE)) { print_complete_net_trace(trace, getEchoFileName(E_ECHO_COMPLETE_NET_TRACE)); } return trace; } /* reload intra cluster nets to complex block */ static void reload_intra_cluster_nets(t_pb *pb) { int i, j; const t_pb_type* pb_type; pb_type = pb->pb_graph_node->pb_type; if (pb_type->blif_model != NULL) { setup_intracluster_routing_for_logical_block(pb->logical_block, pb->pb_graph_node); } else if (pb->child_pbs != NULL) { set_pb_graph_mode(pb->pb_graph_node, pb->mode, 1); for (i = 0; i < pb_type->modes[pb->mode].num_pb_type_children; i++) { for (j = 0; j < pb_type->modes[pb->mode].pb_type_children[i].num_pb; j++) { if (pb->child_pbs[i] != NULL) { if (pb->child_pbs[i][j].name != NULL) { reload_intra_cluster_nets(&pb->child_pbs[i][j]); } } } } } } /* Determine trace from logical_block output to logical_block inputs Algorithm traverses intra-block routing, goes to inter-block routing, then returns to intra-block routing */ static t_trace *alloc_and_load_final_routing_trace() { int i; int iblock; t_trace* final_routing_trace; t_pb_graph_pin *pin; final_routing_trace = (t_trace*) my_calloc(num_logical_nets, sizeof(t_trace)); for (i = 0; i < num_logical_nets; i++) { iblock = logical_block[vpack_net[i].node_block[0]].clb_index; final_routing_trace[i].iblock = iblock; final_routing_trace[i].iswitch = OPEN; final_routing_trace[i].index = OPEN; final_routing_trace[i].next = NULL; pin = get_pb_graph_node_pin_from_vpack_net(i, 0); if (!pin) continue; final_routing_trace[i].index = pin->pin_count_in_cluster; expand_routing_trace(&final_routing_trace[i], i); } return final_routing_trace; } /* Given a routing trace, expand until full trace is complete returns pointer to last terminal trace */ static t_trace *expand_routing_trace(t_trace *trace, int ivpack_net) { int i, iblock, inode, ipin, inet; int gridx, gridy; t_trace *current, *new_trace, *inter_cb_trace; t_rr_node *local_rr_graph; boolean success; t_pb_graph_pin *pb_graph_pin; iblock = trace->iblock; inode = trace->index; local_rr_graph = block[iblock].pb->rr_graph; current = trace; if (local_rr_graph[inode].pb_graph_pin->num_output_edges == 0) { if (local_rr_graph[inode].pb_graph_pin->port->type == OUT_PORT) { /* connection to outside cb */ if (vpack_net[ivpack_net].is_global) { inet = vpack_to_clb_net_mapping[ivpack_net]; if (inet != OPEN) { for (ipin = 1; ipin <= clb_net[inet].num_sinks; ipin++) { pb_graph_pin = get_pb_graph_node_pin_from_clb_net(inet, ipin); new_trace = (t_trace*) my_calloc(1, sizeof(t_trace)); new_trace->iblock = clb_net[inet].node_block[ipin]; new_trace->index = pb_graph_pin->pin_count_in_cluster; new_trace->iswitch = OPEN; new_trace->num_siblings = 0; new_trace->next = NULL; current->next = new_trace; current = expand_routing_trace(new_trace, ivpack_net); } } } else { inter_cb_trace = trace_head[vpack_to_clb_net_mapping[ivpack_net]]; if (inter_cb_trace != NULL) { inter_cb_trace = inter_cb_trace->next; /* skip source and go right to opin */ } while (inter_cb_trace != NULL) { /* continue traversing inter cb trace */ if (rr_node[inter_cb_trace->index].type != SINK) { new_trace = (t_trace*) my_calloc(1, sizeof(t_trace)); new_trace->iblock = OPEN; new_trace->index = inter_cb_trace->index; new_trace->iswitch = inter_cb_trace->iswitch; new_trace->num_siblings = 0; new_trace->next = NULL; current->next = new_trace; if (rr_node[inter_cb_trace->index].type == IPIN) { current = current->next; gridx = rr_node[new_trace->index].xlow; gridy = rr_node[new_trace->index].ylow; gridy = gridy - grid[gridx][gridy].offset; new_trace = (t_trace*) my_calloc(1, sizeof(t_trace)); new_trace->iblock = grid[gridx][gridy].blocks[rr_node[inter_cb_trace->index].z]; new_trace->index = rr_node[inter_cb_trace->index].pb_graph_pin->pin_count_in_cluster; new_trace->iswitch = OPEN; new_trace->num_siblings = 0; new_trace->next = NULL; current->next = new_trace; current = expand_routing_trace(new_trace, ivpack_net); } else { current = current->next; } } inter_cb_trace = inter_cb_trace->next; } } } } else { /* connection to another intra-cluster pin */ current = trace; success = FALSE; for (i = 0; i < local_rr_graph[inode].num_edges; i++) { if (local_rr_graph[local_rr_graph[inode].edges[i]].prev_node == inode) { if (success == FALSE) { success = TRUE; } else { current->next = (t_trace*) my_calloc(1, sizeof(t_trace)); current = current->next; current->iblock = trace->iblock; current->index = trace->index; current->iswitch = trace->iswitch; current->next = NULL; } new_trace = (t_trace*) my_calloc(1, sizeof(t_trace)); new_trace->iblock = trace->iblock; new_trace->index = local_rr_graph[inode].edges[i]; new_trace->iswitch = OPEN; new_trace->num_siblings = 0; new_trace->next = NULL; current->next = new_trace; current = expand_routing_trace(new_trace, ivpack_net); } } assert(success); } return current; } static void print_complete_net_trace(t_trace* trace, const char *file_name) { FILE *fp; int iblock, inode, iprev_block; t_trace *current; t_rr_node *local_rr_graph; const char *name_type[] = { "SOURCE", "SINK", "IPIN", "OPIN", "CHANX", "CHANY", "INTRA_CLUSTER_EDGE" }; int i; fp = my_fopen(file_name, "w", 0); for (i = 0; i < num_logical_nets; i++) { current = &trace[i]; iprev_block = OPEN; fprintf(fp, "Net %s (%d)\n\n", vpack_net[i].name, i); while (current != NULL) { iblock = current->iblock; inode = current->index; if (iblock != OPEN) { if (iprev_block != iblock) { iprev_block = iblock; fprintf(fp, "Block %s (%d) (%d, %d, %d):\n", block[iblock].name, iblock, block[iblock].x, block[iblock].y, block[iblock].z); } local_rr_graph = block[iblock].pb->rr_graph; fprintf(fp, "\tNode:\t%d\t%s[%d].%s[%d]", inode, local_rr_graph[inode].pb_graph_pin->parent_node->pb_type->name, local_rr_graph[inode].pb_graph_pin->parent_node->placement_index, local_rr_graph[inode].pb_graph_pin->port->name, local_rr_graph[inode].pb_graph_pin->pin_number); } else { fprintf(fp, "Node:\t%d\t%6s (%d,%d) ", inode, name_type[(int) rr_node[inode].type], rr_node[inode].xlow, rr_node[inode].ylow); if ((rr_node[inode].xlow != rr_node[inode].xhigh) || (rr_node[inode].ylow != rr_node[inode].yhigh)) fprintf(fp, "to (%d,%d) ", rr_node[inode].xhigh, rr_node[inode].yhigh); switch (rr_node[inode].type) { case IPIN: case OPIN: if (grid[rr_node[inode].xlow][rr_node[inode].ylow].type == IO_TYPE) { fprintf(fp, " Pad: "); } else { /* IO Pad. */ fprintf(fp, " Pin: "); } break; case CHANX: case CHANY: fprintf(fp, " Track: "); break; case SOURCE: case SINK: if (grid[rr_node[inode].xlow][rr_node[inode].ylow].type == IO_TYPE) { fprintf(fp, " Pad: "); } else { /* IO Pad. */ fprintf(fp, " Class: "); } break; default: vpr_printf(TIO_MESSAGE_ERROR, "in print_route: Unexpected traceback element type: %d (%s).\n", rr_node[inode].type, name_type[rr_node[inode].type]); exit(1); break; } fprintf(fp, "%d ", rr_node[inode].ptc_num); /* Uncomment line below if you're debugging and want to see the switch types * * used in the routing. */ /* fprintf (fp, "Switch: %d", tptr->iswitch); */ fprintf(fp, "\n"); } current = current->next; } fprintf(fp, "\n"); } fclose(fp); } void resync_post_route_netlist() { int i, j, iblock; int gridx, gridy; t_trace *trace; for (i = 0; i < num_blocks; i++) { for (j = 0; j < block[i].type->num_pins; j++) { if (block[i].nets[j] != OPEN && clb_net[block[i].nets[j]].is_global == FALSE) block[i].nets[j] = OPEN; } } for (i = 0; i < num_nets; i++) { if (clb_net[i].is_global == TRUE) continue; j = 0; trace = trace_head[i]; while (trace != NULL) { if (rr_node[trace->index].type == OPIN && j == 0) { gridx = rr_node[trace->index].xlow; gridy = rr_node[trace->index].ylow; gridy = gridy - grid[gridx][gridy].offset; iblock = grid[gridx][gridy].blocks[rr_node[trace->index].z]; assert(clb_net[i].node_block[j] == iblock); clb_net[i].node_block_pin[j] = rr_node[trace->index].ptc_num; block[iblock].nets[rr_node[trace->index].ptc_num] = i; j++; } else if (rr_node[trace->index].type == IPIN) { gridx = rr_node[trace->index].xlow; gridy = rr_node[trace->index].ylow; gridy = gridy - grid[gridx][gridy].offset; iblock = grid[gridx][gridy].blocks[rr_node[trace->index].z]; clb_net[i].node_block[j] = iblock; clb_net[i].node_block_pin[j] = rr_node[trace->index].ptc_num; block[iblock].nets[rr_node[trace->index].ptc_num] = i; j++; } trace = trace->next; } assert(j == clb_net[i].num_sinks + 1); } } static void clay_logical_equivalence_handling(const t_arch *arch) { t_trace **saved_ext_rr_trace_head, **saved_ext_rr_trace_tail; t_rr_node *saved_ext_rr_node; int num_ext_rr_node, num_ext_nets; int i, j; for (i = 0; i < num_blocks; i++) { clay_reload_ble_locations(i); } /* Resolve logically equivalent inputs */ saved_ext_rr_trace_head = trace_head; saved_ext_rr_trace_tail = trace_tail; saved_ext_rr_node = rr_node; num_ext_rr_node = num_rr_nodes; num_ext_nets = num_nets; num_rr_nodes = 0; rr_node = NULL; trace_head = NULL; trace_tail = NULL; free_rr_graph(); /* free all data structures associated with rr_graph */ alloc_and_load_cluster_legality_checker(); for (i = 0; i < num_blocks; i++) { /* Regenerate rr_graph (note, can be more runtime efficient but this allows for more code reuse) */ rr_node = block[i].pb->rr_graph; num_rr_nodes = block[i].pb->pb_graph_node->total_pb_pins; free_legalizer_for_cluster(&block[i], TRUE); alloc_and_load_legalizer_for_cluster(&block[i], i, arch); reload_intra_cluster_nets(block[i].pb); reload_ext_net_rr_terminal_cluster(); force_post_place_route_cb_input_pins(i); #ifdef HACK_LUT_PIN_SWAPPING /* Resolve rebalancing of LUT inputs */ clay_lut_input_rebalancing(i, block[i].pb); #endif /* reset rr_graph */ for (j = 0; j < num_rr_nodes; j++) { rr_node[j].occ = 0; rr_node[j].prev_edge = OPEN; rr_node[j].prev_node = OPEN; } if (try_breadth_first_route_cluster() == FALSE) { vpr_printf(TIO_MESSAGE_ERROR, "Failed to resync post routed solution with clustered netlist.\n"); vpr_printf(TIO_MESSAGE_ERROR, "Cannot recover from error.\n"); exit(1); } save_cluster_solution(); reset_legalizer_for_cluster(&block[i]); free_legalizer_for_cluster(&block[i], FALSE); } free_cluster_legality_checker(); trace_head = saved_ext_rr_trace_head; trace_tail = saved_ext_rr_trace_tail; rr_node = saved_ext_rr_node; num_rr_nodes = num_ext_rr_node; num_nets = num_ext_nets; } /* Force router to use the LUT inputs designated by the timing engine post the LUT input rebalancing optimization */ static void clay_lut_input_rebalancing(int iblock, t_pb *pb) { int i, j; t_rr_node *local_rr_graph; t_pb_graph_node *lut_wrapper, *lut; int lut_size; int *lut_pin_remap; int snode, input; t_pb_graph_node *pb_graph_node; if (pb->name != NULL) { pb_graph_node = pb->pb_graph_node; if (pb_graph_node->pb_type->blif_model != NULL) { lut_pin_remap = pb->lut_pin_remap; if (lut_pin_remap != NULL) { local_rr_graph = block[iblock].pb->rr_graph; lut = pb->pb_graph_node; lut_wrapper = lut->parent_pb_graph_node; /* Ensure that this is actually a LUT */ assert( lut->num_input_ports == 1 && lut_wrapper->num_input_ports == 1); assert( lut->num_input_pins[0] == lut_wrapper->num_input_pins[0]); assert( lut->num_output_ports == 1 && lut_wrapper->num_output_ports == 1); assert( lut->num_output_pins[0] == 1 && lut_wrapper->num_output_pins[0] == 1); lut_size = lut->num_input_pins[0]; for (i = 0; i < lut_size; i++) { snode = lut_wrapper->input_pins[0][i].pin_count_in_cluster; free(local_rr_graph[snode].edges); local_rr_graph[snode].edges = NULL; local_rr_graph[snode].num_edges = 0; } for (i = 0; i < lut_size; i++) { input = lut_pin_remap[i]; if (input != OPEN) { snode = lut_wrapper->input_pins[0][i].pin_count_in_cluster; assert(local_rr_graph[snode].num_edges == 0); local_rr_graph[snode].num_edges = 1; local_rr_graph[snode].edges = (int*) my_malloc( sizeof(int)); local_rr_graph[snode].edges[0] = lut->input_pins[0][input].pin_count_in_cluster; } } } } else if (pb->child_pbs != NULL) { for (i = 0; i < pb_graph_node->pb_type->modes[pb->mode].num_pb_type_children; i++) { if (pb->child_pbs[i] != NULL) { for (j = 0; j < pb_graph_node->pb_type->modes[pb->mode].pb_type_children[i].num_pb; j++) { clay_lut_input_rebalancing(iblock, &pb->child_pbs[i][j]); } } } } } } /* Swaps BLEs to match output logical equivalence solution from routing solution Assumes classical cluster with full crossbar and BLEs, each BLE is a single LUT+FF pair */ static void clay_reload_ble_locations(int iblock) { int i, mode, ipin, new_loc; t_pb_graph_node *pb_graph_node; t_pb_graph_pin *pb_graph_pin; const t_pb_type *pb_type; t_trace *trace; t_rr_node *local_rr_graph; int inet, ivpack_net; if (block[iblock].type == IO_TYPE) { return; } pb_graph_node = block[iblock].pb->pb_graph_node; pb_type = pb_graph_node->pb_type; mode = block[iblock].pb->mode; local_rr_graph = block[iblock].pb->rr_graph; assert(block[iblock].pb->mode == 0); assert(pb_type->modes[mode].num_pb_type_children == 1); assert(pb_type->modes[mode].pb_type_children[0].num_output_pins == 1); t_pb** temp; temp = (t_pb**) my_calloc(1, sizeof(t_pb*)); temp[0] = (t_pb*) my_calloc(pb_type->modes[mode].pb_type_children[0].num_pb, sizeof(t_pb)); /* determine new location for BLEs that route out of cluster */ for (i = 0; i < pb_type->modes[mode].pb_type_children[0].num_pb; i++) { if (block[iblock].pb->child_pbs[0][i].name != NULL) { ivpack_net = local_rr_graph[pb_graph_node->child_pb_graph_nodes[mode][0][i].output_pins[0][0].pin_count_in_cluster].net_num; inet = vpack_to_clb_net_mapping[ivpack_net]; if (inet != OPEN) { ipin = OPEN; trace = trace_head[inet]; while (trace) { if (rr_node[trace->index].type == OPIN) { ipin = rr_node[trace->index].ptc_num; break; } trace = trace->next; } assert(ipin); pb_graph_pin = get_pb_graph_node_pin_from_block_pin(iblock, ipin); new_loc = pb_graph_pin->pin_number; assert(temp[0][new_loc].name == NULL); temp[0][new_loc] = block[iblock].pb->child_pbs[0][i]; } } } /* determine new location for BLEs that do not route out of cluster */ new_loc = 0; for (i = 0; i < pb_type->modes[mode].pb_type_children[0].num_pb; i++) { if (block[iblock].pb->child_pbs[0][i].name != NULL) { ivpack_net = local_rr_graph[pb_graph_node->child_pb_graph_nodes[mode][0][i].output_pins[0][0].pin_count_in_cluster].net_num; inet = vpack_to_clb_net_mapping[ivpack_net]; if (inet == OPEN) { while (temp[0][new_loc].name != NULL) { new_loc++; } temp[0][new_loc] = block[iblock].pb->child_pbs[0][i]; } } } free(block[iblock].pb->child_pbs); block[iblock].pb->child_pbs = temp; resync_pb_graph_nodes_in_pb(block[iblock].pb->pb_graph_node, block[iblock].pb); } static void resync_pb_graph_nodes_in_pb(t_pb_graph_node *pb_graph_node, t_pb *pb) { int i, j; if (pb->name == NULL) { return; } assert( strcmp(pb->pb_graph_node->pb_type->name, pb_graph_node->pb_type->name) == 0); pb->pb_graph_node = pb_graph_node; if (pb->child_pbs != NULL) { for (i = 0; i < pb_graph_node->pb_type->modes[pb->mode].num_pb_type_children; i++) { if (pb->child_pbs[i] != NULL) { for (j = 0; j < pb_graph_node->pb_type->modes[pb->mode].pb_type_children[i].num_pb; j++) { resync_pb_graph_nodes_in_pb( &pb_graph_node->child_pb_graph_nodes[pb->mode][i][j], &pb->child_pbs[i][j]); } } } } } /* This function performs power estimation, and must be called * after packing, placement AND routing. Currently, this * will not work when running a partial flow (ex. only routing). */ void vpr_power_estimation(t_vpr_setup vpr_setup, t_arch Arch) { e_power_ret_code power_ret_code; boolean power_error; /* Ensure we are only using 1 clock */ //assert(count_netlist_clocks() == 1); /* Get the critical path of this clock */ g_solution_inf.T_crit = get_critical_path_delay() / 1e9; assert(g_solution_inf.T_crit > 0.); vpr_printf(TIO_MESSAGE_INFO, "\n\nPower Estimation:\n"); vpr_printf(TIO_MESSAGE_INFO, "-----------------\n"); vpr_printf(TIO_MESSAGE_INFO, "Initializing power module\n"); /* Initialize the power module */ power_error = power_init(vpr_setup.FileNameOpts.PowerFile, vpr_setup.FileNameOpts.CmosTechFile, &Arch, &vpr_setup.RoutingArch); if (power_error) { vpr_printf(TIO_MESSAGE_ERROR, "Power initialization failed.\n"); } if (!power_error) { float power_runtime_s; vpr_printf(TIO_MESSAGE_INFO, "Running power estimation\n"); /* Run power estimation */ power_ret_code = power_total(&power_runtime_s, vpr_setup, &Arch, &vpr_setup.RoutingArch); /* Check for errors/warnings */ if (power_ret_code == POWER_RET_CODE_ERRORS) { vpr_printf(TIO_MESSAGE_ERROR, "Power estimation failed. See power output for error details.\n"); } else if (power_ret_code == POWER_RET_CODE_WARNINGS) { vpr_printf(TIO_MESSAGE_WARNING, "Power estimation completed with warnings. See power output for more details.\n"); } else if (power_ret_code == POWER_RET_CODE_SUCCESS) { } vpr_printf(TIO_MESSAGE_INFO, "Power estimation took %g seconds\n", power_runtime_s); } /* Uninitialize power module */ if (!power_error) { vpr_printf(TIO_MESSAGE_INFO, "Uninitializing power module\n"); power_error = power_uninit(); if (power_error) { vpr_printf(TIO_MESSAGE_ERROR, "Power uninitialization failed.\n"); } else { } } vpr_printf(TIO_MESSAGE_INFO, "\n"); }