2019-08-29 23:13:18 -05:00
|
|
|
# = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = =
|
|
|
|
# Script Name : run_fpga_flow.py
|
|
|
|
# Description : This script designed to run different flows supported by
|
|
|
|
# OpensFPGA project.
|
|
|
|
# Args : python3 run_fpga_flow.py --help
|
|
|
|
# Author : Ganesh Gore
|
|
|
|
# Email : ganeshgore@utah.edu
|
|
|
|
# = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = =
|
|
|
|
|
2019-08-09 17:49:05 -05:00
|
|
|
import os
|
2019-08-16 10:44:50 -05:00
|
|
|
import sys
|
2019-08-09 17:49:05 -05:00
|
|
|
import shutil
|
|
|
|
import time
|
2019-08-22 18:01:38 -05:00
|
|
|
from datetime import timedelta
|
2019-08-09 17:49:05 -05:00
|
|
|
import shlex
|
|
|
|
import glob
|
|
|
|
import argparse
|
|
|
|
from configparser import ConfigParser, ExtendedInterpolation
|
|
|
|
import logging
|
|
|
|
import glob
|
|
|
|
import subprocess
|
|
|
|
import threading
|
|
|
|
from string import Template
|
2019-08-15 15:39:58 -05:00
|
|
|
import re
|
2019-08-09 17:49:05 -05:00
|
|
|
import xml.etree.ElementTree as ET
|
2019-08-22 18:01:38 -05:00
|
|
|
from importlib import util
|
|
|
|
if util.find_spec("humanize"):
|
|
|
|
import humanize
|
2019-08-09 17:49:05 -05:00
|
|
|
|
2019-08-29 23:13:18 -05:00
|
|
|
if sys.version_info[0] < 3:
|
|
|
|
raise Exception("run_fpga_task script must be using Python 3")
|
|
|
|
|
2019-08-15 15:39:58 -05:00
|
|
|
# = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = =
|
|
|
|
# Initialise general paths for the script
|
|
|
|
# = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = =
|
|
|
|
# Copy directory where flow file exist
|
2019-08-09 17:49:05 -05:00
|
|
|
flow_script_dir = os.path.dirname(os.path.abspath(__file__))
|
2019-08-15 15:39:58 -05:00
|
|
|
# Find OpenFPGA base directory
|
2019-08-09 17:49:05 -05:00
|
|
|
openfpga_base_dir = os.path.abspath(
|
|
|
|
os.path.join(flow_script_dir, os.pardir, os.pardir))
|
2019-08-15 15:39:58 -05:00
|
|
|
# Copy directory from where script is laucnhed
|
|
|
|
# [req to resolve relative paths provided while launching script]
|
2019-08-09 17:49:05 -05:00
|
|
|
launch_dir = os.getcwd()
|
|
|
|
|
2019-08-15 15:39:58 -05:00
|
|
|
# Path section to append in configuration file to interpolate path
|
2019-08-31 22:55:32 -05:00
|
|
|
task_script_dir = os.path.dirname(os.path.abspath(__file__))
|
2019-08-31 16:19:34 -05:00
|
|
|
script_env_vars = ({"PATH": {
|
|
|
|
"OPENFPGA_FLOW_PATH": task_script_dir,
|
|
|
|
"ARCH_PATH": os.path.join("${PATH:OPENFPGA_PATH}", "arch"),
|
|
|
|
"BENCH_PATH": os.path.join("${PATH:OPENFPGA_PATH}", "benchmarks"),
|
|
|
|
"TECH_PATH": os.path.join("${PATH:OPENFPGA_PATH}", "tech"),
|
|
|
|
"SPICENETLIST_PATH": os.path.join("${PATH:OPENFPGA_PATH}", "SpiceNetlists"),
|
|
|
|
"VERILOG_PATH": os.path.join("${PATH:OPENFPGA_PATH}", "VerilogNetlists"),
|
|
|
|
"OPENFPGA_PATH": os.path.abspath(os.path.join(task_script_dir, os.pardir,
|
|
|
|
os.pardir))}})
|
2019-08-09 17:49:05 -05:00
|
|
|
|
|
|
|
# = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = =
|
2019-08-15 15:39:58 -05:00
|
|
|
# Reading command-line argument
|
2019-08-09 17:49:05 -05:00
|
|
|
# = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = =
|
2019-08-15 15:39:58 -05:00
|
|
|
|
|
|
|
# Helper function to provide better alignment to help print
|
|
|
|
|
|
|
|
|
|
|
|
def formatter(prog): return argparse.HelpFormatter(prog, max_help_position=60)
|
|
|
|
|
|
|
|
|
|
|
|
parser = argparse.ArgumentParser(formatter_class=formatter)
|
|
|
|
|
|
|
|
# Mandatory arguments
|
2019-08-09 17:49:05 -05:00
|
|
|
parser.add_argument('arch_file', type=str)
|
|
|
|
parser.add_argument('benchmark_files', type=str, nargs='+')
|
2019-08-15 15:39:58 -05:00
|
|
|
|
|
|
|
# Optional arguments
|
|
|
|
parser.add_argument('--top_module', type=str, default="top")
|
2019-08-09 17:49:05 -05:00
|
|
|
parser.add_argument('--fpga_flow', type=str, default="yosys_vpr")
|
2019-08-31 16:19:34 -05:00
|
|
|
parser.add_argument('--flow_config', type=str,
|
2019-08-15 15:39:58 -05:00
|
|
|
help="CAD tools path overrides default setting")
|
|
|
|
parser.add_argument('--run_dir', type=str,
|
2019-08-09 17:49:05 -05:00
|
|
|
default=os.path.join(openfpga_base_dir, 'tmp'),
|
|
|
|
help="Directory to store intermidiate file & final results")
|
2019-08-15 15:39:58 -05:00
|
|
|
parser.add_argument('--yosys_tmpl', type=str,
|
|
|
|
help="Alternate yosys template, generates top_module.blif")
|
2019-08-21 12:08:13 -05:00
|
|
|
parser.add_argument('--debug', action="store_true",
|
|
|
|
help="Run script in debug mode")
|
2019-08-15 15:39:58 -05:00
|
|
|
|
2019-08-22 18:00:59 -05:00
|
|
|
# Blif_VPR Only flow arguments
|
|
|
|
parser.add_argument('--activity_file', type=str,
|
|
|
|
help="Activity file used while running yosys flow")
|
|
|
|
parser.add_argument('--base_verilog', type=str,
|
|
|
|
help="Original Verilog file to run verification in " +
|
|
|
|
"blif_VPR flow")
|
|
|
|
|
2019-08-15 15:39:58 -05:00
|
|
|
# ACE2 and power estimation related arguments
|
|
|
|
parser.add_argument('--K', type=int,
|
|
|
|
help="LUT Size, if not specified extracted from arch file")
|
|
|
|
parser.add_argument('--power', action='store_true')
|
|
|
|
parser.add_argument('--power_tech', type=str,
|
|
|
|
help="Power tech xml file for power calculation")
|
|
|
|
parser.add_argument('--ace_d', type=float,
|
|
|
|
help="Specify the default signal density of PIs in ACE2")
|
|
|
|
parser.add_argument('--ace_p', type=float,
|
|
|
|
help="Specify the default signal probablity of PIs in ACE2")
|
|
|
|
parser.add_argument('--black_box_ace', action='store_true')
|
|
|
|
|
|
|
|
# VPR Options
|
2019-08-16 14:36:39 -05:00
|
|
|
parser.add_argument('--min_route_chan_width', type=float,
|
2019-08-15 15:39:58 -05:00
|
|
|
help="Turn on min_route_chan_width")
|
|
|
|
parser.add_argument('--max_route_width_retry', type=int, default=100,
|
|
|
|
help="Maximum iterations to perform to reroute")
|
|
|
|
parser.add_argument('--fix_route_chan_width', type=int,
|
|
|
|
help="Turn on fix_route_chan_width")
|
|
|
|
parser.add_argument('--vpr_timing_pack_off', action='store_true',
|
|
|
|
help="Turn off the timing-driven pack for vpr")
|
|
|
|
parser.add_argument('--vpr_place_clb_pin_remap', action='store_true',
|
|
|
|
help="Turn on place_clb_pin_remap in VPR")
|
|
|
|
parser.add_argument('--vpr_max_router_iteration', type=int,
|
|
|
|
help="Specify the max router iteration in VPR")
|
|
|
|
parser.add_argument('--vpr_route_breadthfirst', action='store_true',
|
|
|
|
help="Use the breadth-first routing algorithm of VPR")
|
|
|
|
parser.add_argument('--vpr_use_tileable_route_chan_width', action='store_true',
|
|
|
|
help="Turn on the conversion to " +
|
|
|
|
"tileable_route_chan_width in VPR")
|
|
|
|
|
|
|
|
# VPR - FPGA-X2P Extension
|
|
|
|
X2PParse = parser.add_argument_group('VPR-FPGA-X2P Extension')
|
|
|
|
X2PParse.add_argument('--vpr_fpga_x2p_rename_illegal_port', action='store_true',
|
|
|
|
help="Rename illegal ports option of VPR FPGA SPICE")
|
|
|
|
X2PParse.add_argument('--vpr_fpga_x2p_signal_density_weight', type=float,
|
|
|
|
help="Specify the signal_density_weight of VPR FPGA SPICE")
|
|
|
|
X2PParse.add_argument('--vpr_fpga_x2p_sim_window_size', type=float,
|
|
|
|
help="specify the sim_window_size of VPR FPGA SPICE")
|
2019-08-22 18:00:59 -05:00
|
|
|
X2PParse.add_argument('--vpr_fpga_x2p_compact_routing_hierarchy',
|
|
|
|
action="store_true", help="Compact_routing_hierarchy")
|
2019-08-15 15:39:58 -05:00
|
|
|
|
|
|
|
# VPR - FPGA-SPICE Extension
|
|
|
|
SPParse = parser.add_argument_group('FPGA-SPICE Extension')
|
|
|
|
SPParse.add_argument('--vpr_fpga_spice', type=str,
|
|
|
|
help="Print SPICE netlists in VPR")
|
|
|
|
SPParse.add_argument('--vpr_fpga_spice_sim_mt_num', type=int,
|
|
|
|
help="Specify the option sim_mt_num of VPR FPGA SPICE")
|
|
|
|
SPParse.add_argument('--vpr_fpga_spice_print_component_tb', action='store_true',
|
|
|
|
help="Output component-level testbench")
|
|
|
|
SPParse.add_argument('--vpr_fpga_spice_print_grid_tb', action='store_true',
|
|
|
|
help="Output grid-level testbench")
|
|
|
|
SPParse.add_argument('--vpr_fpga_spice_print_top_tb', action='store_true',
|
|
|
|
help="Output full-chip-level testbench")
|
|
|
|
SPParse.add_argument('--vpr_fpga_spice_leakage_only', action='store_true',
|
|
|
|
help="Turn on leakage_only mode in VPR FPGA SPICE")
|
|
|
|
SPParse.add_argument('--vpr_fpga_spice_parasitic_net_estimation_off',
|
|
|
|
action='store_true',
|
|
|
|
help="Turn off parasitic_net_estimation in VPR FPGA SPICE")
|
|
|
|
SPParse.add_argument('--vpr_fpga_spice_testbench_load_extraction_off',
|
|
|
|
action='store_true',
|
|
|
|
help="Turn off testbench_load_extraction in VPR FPGA SPICE")
|
|
|
|
SPParse.add_argument('--vpr_fpga_spice_simulator_path', type=str,
|
|
|
|
help="Specify simulator path")
|
|
|
|
|
|
|
|
# VPR - FPGA-Verilog Extension
|
|
|
|
VeriPar = parser.add_argument_group('FPGA-Verilog Extension')
|
|
|
|
VeriPar.add_argument('--vpr_fpga_verilog', action='store_true',
|
|
|
|
help="Generator verilog of VPR FPGA SPICE")
|
|
|
|
VeriPar.add_argument('--vpr_fpga_verilog_dir', type=str,
|
|
|
|
help="path to store generated verilog files")
|
|
|
|
VeriPar.add_argument('--vpr_fpga_verilog_include_timing', action="store_true",
|
|
|
|
help="Print delay specification in Verilog files")
|
|
|
|
VeriPar.add_argument('--vpr_fpga_verilog_include_signal_init',
|
|
|
|
action="store_true",
|
|
|
|
help="Print signal initialization in Verilog files")
|
|
|
|
VeriPar.add_argument('--vpr_fpga_verilog_print_autocheck_top_testbench',
|
|
|
|
action="store_true", help="Print autochecked top-level " +
|
|
|
|
"testbench for Verilog Generator of VPR FPGA SPICE")
|
|
|
|
VeriPar.add_argument('--vpr_fpga_verilog_formal_verification_top_netlist',
|
|
|
|
action="store_true", help="Print formal top Verilog files")
|
|
|
|
VeriPar.add_argument('--vpr_fpga_verilog_include_icarus_simulator',
|
|
|
|
action="store_true", help="dd syntax and definition" +
|
|
|
|
" required to use Icarus Verilog simulator")
|
|
|
|
VeriPar.add_argument('--vpr_fpga_verilog_print_user_defined_template',
|
|
|
|
action="store_true", help="Unknown parameter")
|
|
|
|
VeriPar.add_argument('--vpr_fpga_verilog_print_report_timing_tcl',
|
|
|
|
action="store_true", help="Generate tcl script useful " +
|
|
|
|
"for timing report generation")
|
|
|
|
VeriPar.add_argument('--vpr_fpga_verilog_report_timing_rpt_path',
|
|
|
|
type=str, help="Specify path for report timing results")
|
|
|
|
VeriPar.add_argument('--vpr_fpga_verilog_print_sdc_pnr', action="store_true",
|
|
|
|
help="Generate sdc file to constraint Hardware P&R")
|
|
|
|
VeriPar.add_argument('--vpr_fpga_verilog_print_sdc_analysis',
|
|
|
|
action="store_true", help="Generate sdc file to do STA")
|
|
|
|
VeriPar.add_argument('--vpr_fpga_verilog_print_top_tb', action="store_true",
|
|
|
|
help="Print top-level testbench for Verilog Generator " +
|
|
|
|
"of VPR FPGA SPICE")
|
|
|
|
VeriPar.add_argument('--vpr_fpga_verilog_print_input_blif_tb',
|
|
|
|
action="store_true", help="Print testbench" +
|
|
|
|
"for input blif file in Verilog Generator")
|
|
|
|
VeriPar.add_argument('--vpr_fpga_verilog_print_modelsim_autodeck', type=str,
|
|
|
|
help="Print modelsim " +
|
|
|
|
"simulation script", metavar="<modelsim.ini_path>")
|
2019-08-22 18:00:59 -05:00
|
|
|
VeriPar.add_argument('--vpr_fpga_verilog_explicit_mapping', action="store_true",
|
|
|
|
help="Explicit Mapping")
|
2019-08-15 15:39:58 -05:00
|
|
|
|
|
|
|
# VPR - FPGA-Bitstream Extension
|
|
|
|
BSparse = parser.add_argument_group('FPGA-Bitstream Extension')
|
|
|
|
BSparse.add_argument('--vpr_fpga_bitstream_generator', action="store_true",
|
|
|
|
help="Generate FPGA-SPICE bitstream")
|
|
|
|
|
|
|
|
# Regression test option
|
|
|
|
RegParse = parser.add_argument_group('Regression Test Extension')
|
|
|
|
RegParse.add_argument("--end_flow_with_test", action="store_true",
|
|
|
|
help="Run verification test at the end")
|
2019-08-09 17:49:05 -05:00
|
|
|
|
|
|
|
|
|
|
|
# = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = =
|
2019-08-15 15:39:58 -05:00
|
|
|
# Global varaibles declaration
|
2019-08-09 17:49:05 -05:00
|
|
|
# = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = =
|
2019-08-22 18:01:38 -05:00
|
|
|
# Setting up print and logging system
|
|
|
|
logging.basicConfig(level=logging.INFO, stream=sys.stdout,
|
|
|
|
format='%(levelname)s - %(message)s')
|
|
|
|
logger = logging.getLogger('OpenFPGA_Flow_Logs')
|
|
|
|
|
2019-08-15 15:39:58 -05:00
|
|
|
# variable to store script_configuration and cad tool paths
|
|
|
|
config, cad_tools = None, None
|
2019-08-22 18:01:38 -05:00
|
|
|
ExecTime = {}
|
2019-08-09 17:49:05 -05:00
|
|
|
|
|
|
|
# = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = =
|
|
|
|
# Main program starts here
|
|
|
|
# = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = =
|
2019-08-15 15:39:58 -05:00
|
|
|
|
|
|
|
|
2019-08-09 17:49:05 -05:00
|
|
|
def main():
|
2019-08-16 10:44:50 -05:00
|
|
|
logger.debug("Script Launched in "+os.getcwd())
|
2019-08-15 15:39:58 -05:00
|
|
|
check_required_file()
|
|
|
|
read_script_config()
|
|
|
|
validate_command_line_arguments()
|
2019-08-09 17:49:05 -05:00
|
|
|
prepare_run_directory(args.run_dir)
|
|
|
|
if (args.fpga_flow == "yosys_vpr"):
|
|
|
|
logger.info('Running "yosys_vpr" Flow')
|
|
|
|
run_yosys_with_abc()
|
2019-10-15 17:08:25 -05:00
|
|
|
# TODO Make it optional if activity file is provided
|
|
|
|
run_ace2()
|
|
|
|
run_pro_blif_3arg()
|
2019-08-22 18:01:38 -05:00
|
|
|
if args.power:
|
2019-10-15 17:08:25 -05:00
|
|
|
run_rewrite_verilog()
|
2019-08-22 18:00:59 -05:00
|
|
|
if (args.fpga_flow == "vpr_blif"):
|
|
|
|
collect_files_for_vpr()
|
|
|
|
# if (args.fpga_flow == "vtr"):
|
|
|
|
# run_odin2()
|
|
|
|
# run_abc_vtr()
|
|
|
|
# if (args.fpga_flow == "vtr_standard"):
|
|
|
|
# run_abc_for_standarad()
|
2019-08-09 17:49:05 -05:00
|
|
|
run_vpr()
|
2019-08-15 15:39:58 -05:00
|
|
|
if args.end_flow_with_test:
|
|
|
|
run_netlists_verification()
|
2019-08-22 18:01:38 -05:00
|
|
|
|
|
|
|
ExecTime["End"] = time.time()
|
|
|
|
def timestr(x): return humanize.naturaldelta(timedelta(seconds=x)) \
|
|
|
|
if "humanize" in sys.modules else str(int(x)) + " Sec "
|
|
|
|
TimeInfo = ("Openfpga_flow completed, " +
|
|
|
|
"Total Time Taken %s " %
|
|
|
|
timestr(ExecTime["End"]-ExecTime["Start"]) +
|
|
|
|
"VPR Time %s " %
|
|
|
|
timestr(ExecTime["VPREnd"]-ExecTime["VPRStart"]))
|
|
|
|
TimeInfo += ("Verification Time %s " %
|
|
|
|
timestr(ExecTime["VerificationEnd"] -
|
|
|
|
ExecTime["VerificationStart"])
|
|
|
|
if args.end_flow_with_test else "")
|
|
|
|
logger.info(TimeInfo)
|
2019-08-09 17:49:05 -05:00
|
|
|
exit()
|
|
|
|
|
|
|
|
# = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = =
|
|
|
|
# Subroutines starts here
|
|
|
|
# = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = =
|
|
|
|
|
|
|
|
|
2019-08-15 15:39:58 -05:00
|
|
|
def check_required_file():
|
|
|
|
""" Function ensure existace of all required files for the script """
|
|
|
|
files_dict = {
|
|
|
|
"CAD TOOL PATH": os.path.join(flow_script_dir, os.pardir, 'misc',
|
|
|
|
'fpgaflow_default_tool_path.conf'),
|
|
|
|
}
|
|
|
|
for filename, filepath in files_dict.items():
|
|
|
|
if not os.path.isfile(filepath):
|
|
|
|
clean_up_and_exit("Not able to locate deafult file " + filename)
|
|
|
|
|
|
|
|
|
|
|
|
def read_script_config():
|
|
|
|
""" This fucntion reads default CAD tools path from configuration file """
|
|
|
|
global config, cad_tools
|
|
|
|
config = ConfigParser(interpolation=ExtendedInterpolation())
|
|
|
|
config.read_dict(script_env_vars)
|
|
|
|
default_cad_tool_conf = os.path.join(flow_script_dir, os.pardir, 'misc',
|
|
|
|
'fpgaflow_default_tool_path.conf')
|
|
|
|
config.read_file(open(default_cad_tool_conf))
|
2019-08-31 16:19:34 -05:00
|
|
|
if args.flow_config:
|
|
|
|
config.read_file(open(args.flow_config))
|
2019-08-16 10:44:50 -05:00
|
|
|
if not "CAD_TOOLS_PATH" in config.sections():
|
|
|
|
clean_up_and_exit("Missing CAD_TOOLS_PATH in openfpga_flow config")
|
2019-08-15 15:39:58 -05:00
|
|
|
cad_tools = config["CAD_TOOLS_PATH"]
|
|
|
|
|
|
|
|
|
|
|
|
def validate_command_line_arguments():
|
2019-08-09 17:49:05 -05:00
|
|
|
"""
|
|
|
|
TODO :
|
|
|
|
This funtion validates all supplied paramters
|
|
|
|
"""
|
2019-08-19 22:28:23 -05:00
|
|
|
logger.info("Validating commnad line arguments")
|
|
|
|
|
2019-08-21 12:08:13 -05:00
|
|
|
if args.debug:
|
|
|
|
logger.info("Setting loggger in debug mode")
|
|
|
|
logger.setLevel(logging.DEBUG)
|
|
|
|
|
2019-08-19 22:28:23 -05:00
|
|
|
# Check if flow supported
|
|
|
|
if not args.fpga_flow in config.get("FLOW_SCRIPT_CONFIG", "valid_flows"):
|
2019-08-22 18:01:38 -05:00
|
|
|
clean_up_and_exit("%s Flow not supported" % args.fpga_flow)
|
2019-08-19 22:28:23 -05:00
|
|
|
|
|
|
|
# Check if argument list is consistant
|
|
|
|
for eacharg, dependent in config.items("CMD_ARGUMENT_DEPENDANCY"):
|
|
|
|
if getattr(args, eacharg, None):
|
|
|
|
dependent = dependent.split(",")
|
|
|
|
for eachdep in dependent:
|
|
|
|
if not any([getattr(args, i, 0) for i in eachdep.split("|")]):
|
2019-08-22 18:01:38 -05:00
|
|
|
clean_up_and_exit("'%s' argument depends on (%s) argumets" %
|
|
|
|
(eacharg, ", ".join(dependent).replace("|", " or ")))
|
2019-08-09 17:49:05 -05:00
|
|
|
|
|
|
|
# Filter provided architecrue files
|
|
|
|
args.arch_file = os.path.abspath(args.arch_file)
|
|
|
|
if not os.path.isfile(args.arch_file):
|
|
|
|
clean_up_and_exit("Architecure file not found. -%s", args.arch_file)
|
|
|
|
|
|
|
|
# Filter provided benchmark files
|
|
|
|
for index, everyinput in enumerate(args.benchmark_files):
|
|
|
|
args.benchmark_files[index] = os.path.abspath(everyinput)
|
2019-08-19 22:28:23 -05:00
|
|
|
if os.path.isdir(args.benchmark_files[index]):
|
|
|
|
logger.warning("Skipping directory in bench %s" % everyinput)
|
|
|
|
logger.warning("Directory is not support in benchmark list" +
|
2019-08-22 18:01:38 -05:00
|
|
|
"use wildcard pattern to add files")
|
2019-08-19 22:28:23 -05:00
|
|
|
continue
|
2019-08-09 17:49:05 -05:00
|
|
|
for everyfile in glob.glob(args.benchmark_files[index]):
|
|
|
|
if not os.path.isfile(everyfile):
|
|
|
|
clean_up_and_exit(
|
|
|
|
"Failed to copy benchmark file-%s", args.arch_file)
|
2019-08-15 15:39:58 -05:00
|
|
|
|
2019-08-19 22:28:23 -05:00
|
|
|
# Filter provided powertech files
|
2019-08-15 15:39:58 -05:00
|
|
|
if args.power_tech:
|
|
|
|
args.power_tech = os.path.abspath(args.power_tech)
|
|
|
|
if not os.path.isfile(args.power_tech):
|
|
|
|
clean_up_and_exit(
|
|
|
|
"Power Tech file not found. -%s", args.power_tech)
|
2019-08-16 10:44:50 -05:00
|
|
|
|
2019-08-19 22:28:23 -05:00
|
|
|
# Expand run directory to absolute path
|
2019-08-16 10:44:50 -05:00
|
|
|
args.run_dir = os.path.abspath(args.run_dir)
|
2019-08-22 18:00:59 -05:00
|
|
|
if args.activity_file:
|
|
|
|
args.activity_file = os.path.abspath(args.activity_file)
|
|
|
|
if args.base_verilog:
|
|
|
|
args.base_verilog = os.path.abspath(args.base_verilog)
|
2019-08-09 17:49:05 -05:00
|
|
|
|
|
|
|
|
|
|
|
def ask_user_quetion(condition, question):
|
|
|
|
if condition:
|
|
|
|
reply = str(input(question+' (y/n): ')).lower().strip()
|
|
|
|
if reply[:1] in ['n', 'no']:
|
|
|
|
return False
|
|
|
|
elif reply[:1] in ['y', 'yes']:
|
|
|
|
return True
|
|
|
|
else:
|
|
|
|
return ask_user_quetion(question, condition)
|
|
|
|
return True
|
|
|
|
|
|
|
|
|
|
|
|
def prepare_run_directory(run_dir):
|
|
|
|
"""
|
|
|
|
Prepares run directory to run
|
|
|
|
1. Change current directory to run_dir
|
|
|
|
2. Copy architecture XML file to run_dir
|
|
|
|
3. Copy circuit files to run_dir
|
|
|
|
"""
|
|
|
|
logger.info("Run directory : %s" % run_dir)
|
|
|
|
if os.path.isdir(run_dir):
|
|
|
|
no_of_files = len(next(os.walk(run_dir))[2])
|
|
|
|
if not ask_user_quetion((no_of_files > 100),
|
|
|
|
("[run_dir:%s] already exist and contains %d " +
|
|
|
|
"files script will remove all the file, " +
|
|
|
|
"continue? ") % (run_dir, no_of_files)):
|
|
|
|
clean_up_and_exit("Aborted by user")
|
|
|
|
else:
|
|
|
|
shutil.rmtree(run_dir)
|
|
|
|
os.makedirs(run_dir)
|
|
|
|
# Clean run_dir is created change working directory
|
|
|
|
os.chdir(run_dir)
|
|
|
|
|
|
|
|
# Create arch dir in run_dir and copy flattern architecrture file
|
|
|
|
os.mkdir("arch")
|
2019-08-15 15:39:58 -05:00
|
|
|
tmpl = Template(
|
|
|
|
open(args.arch_file, encoding='utf-8').read())
|
2019-08-09 17:49:05 -05:00
|
|
|
arch_filename = os.path.basename(args.arch_file)
|
|
|
|
args.arch_file = os.path.join(run_dir, "arch", arch_filename)
|
2019-08-15 15:39:58 -05:00
|
|
|
with open(args.arch_file, 'w', encoding='utf-8') as archfile:
|
2019-08-09 17:49:05 -05:00
|
|
|
archfile.write(tmpl.substitute(script_env_vars["PATH"]))
|
|
|
|
|
|
|
|
# Create benchmark dir in run_dir and copy flattern architecrture file
|
|
|
|
os.mkdir("benchmark")
|
|
|
|
try:
|
|
|
|
for index, eachfile in enumerate(args.benchmark_files):
|
|
|
|
args.benchmark_files[index] = shutil.copy2(
|
|
|
|
eachfile, os.path.join(os.curdir, "benchmark"))
|
|
|
|
except:
|
|
|
|
logger.exception("Failed to copy all benchmark file to run_dir")
|
|
|
|
|
|
|
|
|
|
|
|
def clean_up_and_exit(msg, clean=False):
|
2019-08-16 10:44:50 -05:00
|
|
|
logger.error("Current working directory : " + os.getcwd())
|
2019-08-09 17:49:05 -05:00
|
|
|
logger.error(msg)
|
2019-08-15 15:39:58 -05:00
|
|
|
logger.error("Exiting . . . . . .")
|
2019-08-18 13:26:05 -05:00
|
|
|
exit(1)
|
2019-08-09 17:49:05 -05:00
|
|
|
|
|
|
|
|
|
|
|
def run_yosys_with_abc():
|
2019-08-15 15:39:58 -05:00
|
|
|
"""
|
|
|
|
Execute yosys with ABC and optional blackbox support
|
|
|
|
"""
|
2019-08-09 17:49:05 -05:00
|
|
|
tree = ET.parse(args.arch_file)
|
|
|
|
root = tree.getroot()
|
|
|
|
try:
|
|
|
|
lut_size = max([int(pb_type.find("input").get("num_pins"))
|
|
|
|
for pb_type in root.iter("pb_type")
|
|
|
|
if pb_type.get("class") == "lut"])
|
2019-08-16 14:36:39 -05:00
|
|
|
logger.info("Extracted lut_size size from arch XML = %s", lut_size)
|
2019-08-09 17:49:05 -05:00
|
|
|
logger.info("Running Yosys with lut_size = %s", lut_size)
|
|
|
|
except:
|
|
|
|
logger.exception("Failed to extract lut_size from XML file")
|
|
|
|
clean_up_and_exit("")
|
2019-08-15 15:39:58 -05:00
|
|
|
args.K = lut_size
|
2019-08-09 17:49:05 -05:00
|
|
|
# Yosys script parameter mapping
|
|
|
|
ys_params = {
|
|
|
|
"READ_VERILOG_FILE": " \n".join([
|
|
|
|
"read_verilog -nolatches " + shlex.quote(eachfile)
|
|
|
|
for eachfile in args.benchmark_files]),
|
|
|
|
"TOP_MODULE": args.top_module,
|
|
|
|
"LUT_SIZE": lut_size,
|
2019-08-19 20:06:46 -05:00
|
|
|
"OUTPUT_BLIF": args.top_module+"_yosys_out.blif",
|
2019-08-09 17:49:05 -05:00
|
|
|
}
|
|
|
|
yosys_template = os.path.join(
|
|
|
|
cad_tools["misc_dir"], "ys_tmpl_yosys_vpr_flow.ys")
|
2019-08-15 15:39:58 -05:00
|
|
|
tmpl = Template(open(yosys_template, encoding='utf-8').read())
|
2019-08-09 17:49:05 -05:00
|
|
|
with open("yosys.ys", 'w') as archfile:
|
|
|
|
archfile.write(tmpl.substitute(ys_params))
|
|
|
|
try:
|
|
|
|
with open('yosys_output.txt', 'w+') as output:
|
2019-08-15 15:39:58 -05:00
|
|
|
process = subprocess.run([cad_tools["yosys_path"], 'yosys.ys'],
|
2019-08-09 17:49:05 -05:00
|
|
|
check=True,
|
|
|
|
stdout=subprocess.PIPE,
|
|
|
|
stderr=subprocess.PIPE,
|
|
|
|
universal_newlines=True)
|
|
|
|
output.write(process.stdout)
|
|
|
|
if process.returncode:
|
|
|
|
logger.info("Yosys failed with returncode %d",
|
|
|
|
process.returncode)
|
2019-09-05 11:23:35 -05:00
|
|
|
raise subprocess.CalledProcessError(0, command)
|
2019-08-09 17:49:05 -05:00
|
|
|
except:
|
|
|
|
logger.exception("Failed to run yosys")
|
|
|
|
clean_up_and_exit("")
|
2019-08-15 15:39:58 -05:00
|
|
|
logger.info("Yosys output is written in file yosys_output.txt")
|
2019-08-09 17:49:05 -05:00
|
|
|
|
|
|
|
|
|
|
|
def run_odin2():
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
|
|
def run_abc_vtr():
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
|
|
def run_abc_for_standarad():
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
|
|
def run_ace2():
|
2019-08-15 15:39:58 -05:00
|
|
|
if args.black_box_ace:
|
2019-08-19 20:06:46 -05:00
|
|
|
with open(args.top_module+'_yosys_out.blif', 'r') as fp:
|
2019-08-15 15:39:58 -05:00
|
|
|
blif_lines = fp.readlines()
|
|
|
|
|
|
|
|
with open(args.top_module+'_bb.blif', 'w') as fp:
|
|
|
|
for eachline in blif_lines:
|
|
|
|
if ".names" in eachline:
|
|
|
|
input_nets = eachline.split()[1:]
|
|
|
|
if len(input_nets)-1 > args.K:
|
|
|
|
logger.error("One module in blif have more inputs" +
|
|
|
|
" than K value")
|
|
|
|
# Map CEll to each logic in blif
|
|
|
|
map_nets = (input_nets[:-1] + ["unconn"]*args.K)[:args.K]
|
|
|
|
map_nets = ["I[%d]=%s" % (indx, eachnet)
|
|
|
|
for indx, eachnet in enumerate(map_nets)]
|
|
|
|
map_nets += ["O[0]=%s\n" % input_nets[-1]]
|
|
|
|
fp.write(".subckt CELL ")
|
|
|
|
fp.write(" ".join(map_nets))
|
|
|
|
else:
|
|
|
|
fp.write(eachline)
|
|
|
|
|
|
|
|
declar_input = " ".join(["I[%d]" % i for i in range(args.K)])
|
|
|
|
model_tmpl = "\n" + \
|
|
|
|
".model CELL\n" + \
|
|
|
|
".inputs " + declar_input + " \n" + \
|
|
|
|
".outputs O[0]\n" + \
|
|
|
|
".blackbox\n" + \
|
|
|
|
".end\n"
|
|
|
|
fp.write(model_tmpl)
|
|
|
|
# Prepare ACE run command
|
|
|
|
command = [
|
2019-08-19 20:06:46 -05:00
|
|
|
"-b", args.top_module +
|
|
|
|
('_bb.blif' if args.black_box_ace else "_yosys_out.blif"),
|
2019-08-15 15:39:58 -05:00
|
|
|
"-o", args.top_module+"_ace_out.act",
|
|
|
|
"-n", args.top_module+"_ace_out.blif",
|
|
|
|
"-c", "clk",
|
|
|
|
]
|
|
|
|
command += ["-d", "%.4f" % args.ace_d] if args.ace_d else [""]
|
|
|
|
command += ["-p", "%.4f" % args.ace_d] if args.ace_p else [""]
|
|
|
|
try:
|
|
|
|
filename = args.top_module + '_ace2_output.txt'
|
|
|
|
with open(filename, 'w+') as output:
|
|
|
|
process = subprocess.run([cad_tools["ace_path"]] + command,
|
|
|
|
check=True,
|
|
|
|
stdout=subprocess.PIPE,
|
|
|
|
stderr=subprocess.PIPE,
|
|
|
|
universal_newlines=True)
|
|
|
|
output.write(process.stdout)
|
|
|
|
if process.returncode:
|
|
|
|
logger.info("ACE2 failed with returncode %d",
|
|
|
|
process.returncode)
|
2019-09-05 11:23:35 -05:00
|
|
|
raise subprocess.CalledProcessError(0, command)
|
2019-08-15 15:39:58 -05:00
|
|
|
except:
|
|
|
|
logger.exception("Failed to run ACE2")
|
|
|
|
clean_up_and_exit("")
|
|
|
|
logger.info("ACE2 output is written in file %s" % filename)
|
2019-08-09 17:49:05 -05:00
|
|
|
|
2019-08-15 15:39:58 -05:00
|
|
|
|
|
|
|
def run_pro_blif_3arg():
|
|
|
|
command = [
|
|
|
|
"-i", args.top_module+"_ace_out.blif",
|
2019-08-19 20:06:46 -05:00
|
|
|
"-o", args.top_module+".blif",
|
|
|
|
"-initial_blif", args.top_module+'_yosys_out.blif',
|
2019-08-15 15:39:58 -05:00
|
|
|
]
|
|
|
|
try:
|
|
|
|
filename = args.top_module+'_blif_3args_output.txt'
|
|
|
|
with open(filename, 'w+') as output:
|
|
|
|
process = subprocess.run(["perl", cad_tools["pro_blif_path"]] +
|
|
|
|
command,
|
|
|
|
check=True,
|
|
|
|
stdout=subprocess.PIPE,
|
|
|
|
stderr=subprocess.PIPE,
|
|
|
|
universal_newlines=True)
|
|
|
|
output.write(process.stdout)
|
|
|
|
if process.returncode:
|
|
|
|
logger.info("blif_3args script failed with returncode %d",
|
|
|
|
process.returncode)
|
|
|
|
except:
|
|
|
|
logger.exception("Failed to run blif_3args")
|
|
|
|
clean_up_and_exit("")
|
|
|
|
logger.info("blif_3args output is written in file %s" % filename)
|
2019-08-09 17:49:05 -05:00
|
|
|
|
|
|
|
|
2019-08-22 18:01:38 -05:00
|
|
|
def collect_files_for_vpr():
|
2019-08-25 01:23:39 -05:00
|
|
|
# Sanitize provided Benchmark option
|
2019-08-22 18:01:38 -05:00
|
|
|
if len(args.benchmark_files) > 1:
|
|
|
|
logger.error("Expecting Single Benchmark BLif file.")
|
2019-08-25 01:23:39 -05:00
|
|
|
if not os.path.isfile(args.benchmark_files[0] or ""):
|
|
|
|
clean_up_and_exit("Provided Blif file not found")
|
2019-08-22 18:01:38 -05:00
|
|
|
shutil.copy(args.benchmark_files[0], args.top_module+".blif")
|
2019-08-25 01:23:39 -05:00
|
|
|
|
|
|
|
# Sanitize provided Activity file option
|
|
|
|
if not os.path.isfile(args.activity_file or ""):
|
|
|
|
logger.error("Activity File - %s" % args.activity_file)
|
|
|
|
clean_up_and_exit("Provided activity file not found")
|
2019-08-22 18:01:38 -05:00
|
|
|
shutil.copy(args.activity_file, args.top_module+"_ace_out.act")
|
2019-08-25 01:23:39 -05:00
|
|
|
|
|
|
|
# Sanitize provided Benchmark option
|
|
|
|
if not os.path.isfile(args.base_verilog or ""):
|
|
|
|
logger.error("Base Verilog File - %s" % args.base_verilog)
|
|
|
|
clean_up_and_exit("Provided base_verilog file not found")
|
2019-08-22 18:01:38 -05:00
|
|
|
shutil.copy(args.base_verilog, args.top_module+"_output_verilog.v")
|
|
|
|
|
|
|
|
|
2019-08-09 17:49:05 -05:00
|
|
|
def run_vpr():
|
2019-08-22 18:01:38 -05:00
|
|
|
ExecTime["VPRStart"] = time.time()
|
|
|
|
|
2019-08-22 18:00:59 -05:00
|
|
|
if not args.fix_route_chan_width:
|
2019-08-22 18:01:38 -05:00
|
|
|
# Run Standard VPR Flow
|
|
|
|
min_channel_width = run_standard_vpr(
|
|
|
|
args.top_module+".blif",
|
|
|
|
-1,
|
|
|
|
args.top_module+"_min_chan_width_vpr.txt")
|
|
|
|
logger.info("Standard VPR flow routed with minimum %d Channels" %
|
|
|
|
min_channel_width)
|
2019-08-15 15:39:58 -05:00
|
|
|
|
|
|
|
# Minimum routing channel width
|
|
|
|
if (args.min_route_chan_width):
|
2019-08-16 14:36:39 -05:00
|
|
|
min_channel_width *= args.min_route_chan_width
|
2019-08-15 15:39:58 -05:00
|
|
|
min_channel_width = int(min_channel_width)
|
|
|
|
min_channel_width += 1 if (min_channel_width % 2) else 0
|
2019-08-16 14:36:39 -05:00
|
|
|
logger.info(("Trying to route using %d channels" % min_channel_width) +
|
|
|
|
" (Slack of %d%%)" % ((args.min_route_chan_width-1)*100))
|
2019-08-15 15:39:58 -05:00
|
|
|
|
|
|
|
while(1):
|
2019-08-19 20:06:46 -05:00
|
|
|
res = run_standard_vpr(args.top_module+".blif",
|
2019-08-18 13:26:05 -05:00
|
|
|
int(min_channel_width),
|
2019-08-22 18:01:38 -05:00
|
|
|
args.top_module+"_reroute_vpr.txt",
|
|
|
|
route_only=True)
|
2019-08-15 15:39:58 -05:00
|
|
|
|
|
|
|
if res:
|
|
|
|
logger.info("Routing with channel width=%d successful" %
|
|
|
|
min_channel_width)
|
|
|
|
break
|
2019-08-17 17:22:52 -05:00
|
|
|
elif args.max_route_width_retry < (min_channel_width-2):
|
2019-08-15 15:39:58 -05:00
|
|
|
clean_up_and_exit("Failed to route within maximum " +
|
|
|
|
"iteration of channel width")
|
|
|
|
else:
|
|
|
|
logger.info("Unable to route using channel width %d" %
|
|
|
|
min_channel_width)
|
|
|
|
min_channel_width += 2
|
|
|
|
|
2019-08-19 20:06:46 -05:00
|
|
|
extract_vpr_stats(args.top_module+"_reroute_vpr.txt")
|
2019-08-15 15:39:58 -05:00
|
|
|
|
|
|
|
# Fixed routing channel width
|
|
|
|
elif args.fix_route_chan_width:
|
|
|
|
min_channel_width = run_standard_vpr(
|
2019-08-19 20:06:46 -05:00
|
|
|
args.top_module+".blif",
|
2019-08-15 15:39:58 -05:00
|
|
|
args.fix_route_chan_width,
|
2019-10-15 17:07:34 -05:00
|
|
|
args.top_module+"_fr_chan_width_vpr.txt")
|
2019-08-15 15:39:58 -05:00
|
|
|
logger.info("Fixed routing channel successfully routed with %d width" %
|
|
|
|
min_channel_width)
|
2019-11-01 21:51:05 -05:00
|
|
|
extract_vpr_stats(args.top_module+"_fr_chan_width_vpr.txt")
|
2019-08-15 15:39:58 -05:00
|
|
|
else:
|
|
|
|
extract_vpr_stats(args.top_module+"_min_chan_width.txt")
|
|
|
|
if args.power:
|
2019-08-19 20:06:46 -05:00
|
|
|
extract_vpr_stats(logfile=args.top_module+".power",
|
|
|
|
r_filename="vpr_power_stat",
|
|
|
|
parse_section="power")
|
2019-08-22 18:01:38 -05:00
|
|
|
ExecTime["VPREnd"] = time.time()
|
2019-08-15 15:39:58 -05:00
|
|
|
|
|
|
|
|
2019-08-18 13:26:05 -05:00
|
|
|
def run_standard_vpr(bench_blif, fixed_chan_width, logfile, route_only=False):
|
2019-08-15 15:39:58 -05:00
|
|
|
command = [cad_tools["vpr_path"],
|
|
|
|
args.arch_file,
|
|
|
|
bench_blif,
|
|
|
|
"--net_file", args.top_module+"_vpr.net",
|
|
|
|
"--place_file", args.top_module+"_vpr.place",
|
|
|
|
"--route_file", args.top_module+"_vpr.route",
|
2019-08-18 13:26:05 -05:00
|
|
|
"--full_stats", "--nodisp",
|
2019-10-15 17:08:25 -05:00
|
|
|
"--activity_file", args.top_module+"_ace_out.act",
|
2019-08-15 15:39:58 -05:00
|
|
|
]
|
2019-08-18 13:26:05 -05:00
|
|
|
if route_only:
|
|
|
|
command += ["--route"]
|
2019-08-15 15:39:58 -05:00
|
|
|
# Power options
|
|
|
|
if args.power:
|
|
|
|
command += ["--power",
|
|
|
|
"--tech_properties", args.power_tech]
|
|
|
|
# packer options
|
|
|
|
if args.vpr_timing_pack_off:
|
|
|
|
command += ["--timing_driven_clustering", "off"]
|
|
|
|
# channel width option
|
|
|
|
if fixed_chan_width >= 0:
|
2019-08-22 18:00:59 -05:00
|
|
|
command += ["--route_chan_width", "%d" % fixed_chan_width]
|
2019-08-15 15:39:58 -05:00
|
|
|
if args.vpr_use_tileable_route_chan_width:
|
|
|
|
command += ["--use_tileable_route_chan_width"]
|
|
|
|
|
2019-09-16 22:22:13 -05:00
|
|
|
if args.vpr_fpga_x2p_compact_routing_hierarchy:
|
|
|
|
command += ["--fpga_x2p_compact_routing_hierarchy"]
|
|
|
|
|
2019-08-15 15:39:58 -05:00
|
|
|
# FPGA_Spice Options
|
|
|
|
if (args.power and args.vpr_fpga_spice):
|
|
|
|
command += ["--fpga_spice"]
|
|
|
|
if args.vpr_fpga_x2p_signal_density_weight:
|
|
|
|
command += ["--fpga_x2p_signal_density_weight",
|
|
|
|
args.vpr_fpga_x2p_signal_density_weight]
|
|
|
|
if args.vpr_fpga_x2p_sim_window_size:
|
|
|
|
command += ["--fpga_x2p_sim_window_size",
|
|
|
|
args.vpr_fpga_x2p_sim_window_size]
|
2019-08-22 18:00:59 -05:00
|
|
|
|
2019-08-15 15:39:58 -05:00
|
|
|
if args.vpr_fpga_spice_sim_mt_num:
|
|
|
|
command += ["--fpga_spice_sim_mt_num",
|
|
|
|
args.vpr_fpga_spice_sim_mt_num]
|
|
|
|
if args.vpr_fpga_spice_simulator_path:
|
|
|
|
command += ["--fpga_spice_simulator_path",
|
|
|
|
args.vpr_fpga_spice_simulator_path]
|
|
|
|
if args.vpr_fpga_spice_print_component_tb:
|
|
|
|
command += ["--fpga_spice_print_lut_testbench",
|
|
|
|
"--fpga_spice_print_hardlogic_testbench",
|
|
|
|
"--fpga_spice_print_pb_mux_testbench",
|
|
|
|
"--fpga_spice_print_cb_mux_testbench",
|
|
|
|
"--fpga_spice_print_sb_mux_testbench"
|
|
|
|
]
|
|
|
|
if args.vpr_fpga_spice_print_grid_tb:
|
|
|
|
command += ["--fpga_spice_print_grid_testbench",
|
|
|
|
"--fpga_spice_print_cb_testbench",
|
|
|
|
"--fpga_spice_print_sb_testbench"
|
|
|
|
]
|
|
|
|
if args.vpr_fpga_spice_print_top_tb:
|
|
|
|
command += ["--fpga_spice_print_top_testbench"]
|
|
|
|
if args.vpr_fpga_spice_leakage_only:
|
|
|
|
command += ["--fpga_spice_leakage_only"]
|
|
|
|
if args.vpr_fpga_spice_parasitic_net_estimation_off:
|
|
|
|
command += ["--fpga_spice_parasitic_net_estimation", "off"]
|
|
|
|
if args.vpr_fpga_spice_testbench_load_extraction_off:
|
|
|
|
command += ["--fpga_spice_testbench_load_extraction", "off"]
|
|
|
|
|
|
|
|
# FPGA Verilog options
|
2019-10-15 17:08:25 -05:00
|
|
|
if args.vpr_fpga_verilog:
|
2019-08-15 15:39:58 -05:00
|
|
|
command += ["--fpga_verilog"]
|
|
|
|
if args.vpr_fpga_verilog_dir:
|
|
|
|
command += ["--fpga_verilog_dir", args.vpr_fpga_verilog_dir]
|
|
|
|
if args.vpr_fpga_verilog_print_top_tb:
|
|
|
|
command += ["--fpga_verilog_print_top_testbench"]
|
|
|
|
if args.vpr_fpga_verilog_print_input_blif_tb:
|
|
|
|
command += ["--fpga_verilog_print_input_blif_testbench"]
|
|
|
|
if args.vpr_fpga_verilog_print_autocheck_top_testbench:
|
|
|
|
command += ["--fpga_verilog_print_autocheck_top_testbench",
|
2019-11-01 19:21:26 -05:00
|
|
|
# args.vpr_fpga_verilog_print_autocheck_top_testbench]
|
|
|
|
os.path.join(args.run_dir, args.top_module+"_output_verilog.v")]
|
2019-08-15 15:39:58 -05:00
|
|
|
if args.vpr_fpga_verilog_include_timing:
|
|
|
|
command += ["--fpga_verilog_include_timing"]
|
2019-08-22 18:00:59 -05:00
|
|
|
if args.vpr_fpga_verilog_explicit_mapping:
|
|
|
|
command += ["--fpga_verilog_explicit_mapping"]
|
2019-08-15 15:39:58 -05:00
|
|
|
if args.vpr_fpga_verilog_include_signal_init:
|
|
|
|
command += ["--fpga_verilog_include_signal_init"]
|
|
|
|
if args.vpr_fpga_verilog_formal_verification_top_netlist:
|
2019-08-16 14:36:39 -05:00
|
|
|
command += ["--fpga_verilog_print_formal_verification_top_netlist"]
|
2019-08-15 15:39:58 -05:00
|
|
|
if args.vpr_fpga_verilog_print_modelsim_autodeck:
|
|
|
|
command += ["--fpga_verilog_print_modelsim_autodeck",
|
|
|
|
args.vpr_fpga_verilog_print_modelsim_autodeck]
|
|
|
|
if args.vpr_fpga_verilog_include_icarus_simulator:
|
|
|
|
command += ["--fpga_verilog_include_icarus_simulator"]
|
|
|
|
if args.vpr_fpga_verilog_print_report_timing_tcl:
|
|
|
|
command += ["--fpga_verilog_print_report_timing_tcl"]
|
|
|
|
if args.vpr_fpga_verilog_report_timing_rpt_path:
|
|
|
|
command += ["--fpga_verilog_report_timing_rpt_path",
|
|
|
|
args.vpr_fpga_verilog_report_timing_rpt_path]
|
|
|
|
if args.vpr_fpga_verilog_print_sdc_pnr:
|
|
|
|
command += ["--fpga_verilog_print_sdc_pnr"]
|
|
|
|
if args.vpr_fpga_verilog_print_user_defined_template:
|
|
|
|
command += ["--fpga_verilog_print_user_defined_template"]
|
|
|
|
if args.vpr_fpga_verilog_print_sdc_analysis:
|
|
|
|
command += ["--fpga_verilog_print_sdc_analysis"]
|
|
|
|
|
|
|
|
# FPGA Bitstream Genration options
|
|
|
|
if args.vpr_fpga_verilog_print_sdc_analysis:
|
|
|
|
command += ["--fpga_bitstream_generator"]
|
|
|
|
|
|
|
|
if args.vpr_fpga_x2p_rename_illegal_port or \
|
|
|
|
args.vpr_fpga_spice or \
|
|
|
|
args.vpr_fpga_verilog:
|
|
|
|
command += ["--fpga_x2p_rename_illegal_port"]
|
|
|
|
|
|
|
|
# Other VPR options
|
|
|
|
if args.vpr_place_clb_pin_remap:
|
|
|
|
command += ["--place_clb_pin_remap"]
|
|
|
|
if args.vpr_route_breadthfirst:
|
|
|
|
command += ["--router_algorithm", "breadth_first"]
|
|
|
|
if args.vpr_max_router_iteration:
|
|
|
|
command += ["--max_router_iterations", args.vpr_max_router_iteration]
|
|
|
|
|
|
|
|
chan_width = None
|
|
|
|
try:
|
2019-08-22 18:01:38 -05:00
|
|
|
logger.debug("Running VPR : " + " ".join(command))
|
2019-08-15 15:39:58 -05:00
|
|
|
with open(logfile, 'w+') as output:
|
|
|
|
output.write(" ".join(command)+"\n")
|
|
|
|
process = subprocess.run(command,
|
|
|
|
check=True,
|
|
|
|
stdout=subprocess.PIPE,
|
|
|
|
stderr=subprocess.PIPE,
|
|
|
|
universal_newlines=True)
|
|
|
|
for line in process.stdout.split('\n'):
|
|
|
|
if "Best routing" in line:
|
2019-08-18 13:26:05 -05:00
|
|
|
chan_width = int(re.search(
|
|
|
|
r"channel width factor of ([0-9]+)", line).group(1))
|
2019-08-15 15:39:58 -05:00
|
|
|
if "Circuit successfully routed" in line:
|
2019-08-18 13:26:05 -05:00
|
|
|
chan_width = int(re.search(
|
|
|
|
r"a channel width factor of ([0-9]+)", line).group(1))
|
2019-08-15 15:39:58 -05:00
|
|
|
output.write(process.stdout)
|
|
|
|
if process.returncode:
|
|
|
|
logger.info("Standard VPR run failed with returncode %d",
|
|
|
|
process.returncode)
|
2019-09-05 11:23:35 -05:00
|
|
|
raise subprocess.CalledProcessError(0, command)
|
2019-08-18 13:26:05 -05:00
|
|
|
except (Exception, subprocess.CalledProcessError) as e:
|
2019-08-15 15:39:58 -05:00
|
|
|
logger.exception("Failed to run VPR")
|
2019-08-23 00:41:25 -05:00
|
|
|
filter_failed_process_output(e.output)
|
2019-08-15 15:39:58 -05:00
|
|
|
clean_up_and_exit("")
|
|
|
|
logger.info("VPR output is written in file %s" % logfile)
|
|
|
|
return chan_width
|
|
|
|
|
|
|
|
|
2019-08-19 20:04:14 -05:00
|
|
|
def extract_vpr_stats(logfile, r_filename="vpr_stat", parse_section="vpr"):
|
|
|
|
section = "DEFAULT_PARSE_RESULT_POWER" if parse_section == "power" \
|
|
|
|
else "DEFAULT_PARSE_RESULT_VPR"
|
2019-08-15 15:39:58 -05:00
|
|
|
vpr_log = open(logfile).read()
|
|
|
|
resultDict = {}
|
2019-08-19 20:04:14 -05:00
|
|
|
for name, value in config.items(section):
|
|
|
|
reg_string, filt_function = value.split(",")
|
2019-08-15 15:39:58 -05:00
|
|
|
match = re.search(reg_string[1:-1], vpr_log)
|
|
|
|
if match:
|
2019-08-19 20:04:14 -05:00
|
|
|
try:
|
|
|
|
if "lambda" in filt_function.strip():
|
|
|
|
eval("ParseFunction = "+filt_function.strip())
|
|
|
|
extract_val = ParseFunction(**match.groups())
|
|
|
|
elif filt_function.strip() == "int":
|
|
|
|
extract_val = int(match.group(1))
|
|
|
|
elif filt_function.strip() == "float":
|
|
|
|
extract_val = float(match.group(1))
|
|
|
|
elif filt_function.strip() == "str":
|
|
|
|
extract_val = str(match.group(1))
|
|
|
|
elif filt_function.strip() == "scientific":
|
|
|
|
try:
|
2019-08-22 18:01:38 -05:00
|
|
|
mult = {"m": 1E-3, "u": 1E-6, "n": 1E-9,
|
|
|
|
"K": 1E-3, "M": 1E-6, "G": 1E-9, }.get(match.group(2)[0], 1)
|
2019-08-19 20:04:14 -05:00
|
|
|
except:
|
|
|
|
mult = 1
|
|
|
|
extract_val = float(match.group(1))*mult
|
|
|
|
else:
|
2019-08-19 20:06:46 -05:00
|
|
|
extract_val = match.group(1)
|
2019-08-19 20:04:14 -05:00
|
|
|
except:
|
|
|
|
logger.exception("Filter failed")
|
2019-08-22 18:01:38 -05:00
|
|
|
extract_val = "Filter Failed"
|
2019-08-15 15:39:58 -05:00
|
|
|
resultDict[name] = extract_val
|
|
|
|
|
|
|
|
dummyparser = ConfigParser()
|
2019-08-16 11:59:44 -05:00
|
|
|
dummyparser.read_dict({"RESULTS": resultDict})
|
2019-08-15 15:39:58 -05:00
|
|
|
|
2019-08-19 20:04:14 -05:00
|
|
|
with open(r_filename+'.result', 'w') as configfile:
|
2019-08-15 15:39:58 -05:00
|
|
|
dummyparser.write(configfile)
|
2019-08-19 20:04:14 -05:00
|
|
|
logger.info("%s result extracted in file %s" %
|
2019-08-22 18:01:38 -05:00
|
|
|
(parse_section, r_filename+'.result'))
|
2019-08-15 15:39:58 -05:00
|
|
|
|
|
|
|
|
|
|
|
def run_rewrite_verilog():
|
|
|
|
# Rewrite the verilog after optimization
|
|
|
|
script_cmd = [
|
2019-08-19 20:06:46 -05:00
|
|
|
"read_blif %s" % args.top_module+".blif",
|
2019-08-15 15:39:58 -05:00
|
|
|
"write_verilog %s" % args.top_module+"_output_verilog.v"
|
|
|
|
]
|
|
|
|
command = [cad_tools["yosys_path"], "-p", "; ".join(script_cmd)]
|
2019-08-23 00:41:25 -05:00
|
|
|
run_command("Yosys", "yosys_output.txt", command)
|
2019-08-15 15:39:58 -05:00
|
|
|
|
|
|
|
|
2019-10-31 10:56:57 -05:00
|
|
|
def run_netlists_verification(exit_if_fail=True):
|
2019-08-22 18:01:38 -05:00
|
|
|
ExecTime["VerificationStart"] = time.time()
|
2019-08-15 15:39:58 -05:00
|
|
|
compiled_file = "compiled_"+args.top_module
|
|
|
|
# include_netlists = args.top_module+"_include_netlists.v"
|
|
|
|
tb_top_formal = args.top_module+"_top_formal_verification_random_tb"
|
|
|
|
tb_top_autochecked = args.top_module+"_autocheck_top_tb"
|
|
|
|
# netlists_path = args.vpr_fpga_verilog_dir_val+"/SRC/"
|
|
|
|
|
|
|
|
command = [cad_tools["iverilog_path"]]
|
|
|
|
command += ["-o", compiled_file]
|
2019-08-19 20:06:46 -05:00
|
|
|
command += ["./SRC/%s_include_netlists.v" %
|
2019-08-17 17:22:52 -05:00
|
|
|
args.top_module]
|
2019-08-15 15:39:58 -05:00
|
|
|
command += ["-s"]
|
|
|
|
if args.vpr_fpga_verilog_formal_verification_top_netlist:
|
|
|
|
command += [tb_top_formal]
|
|
|
|
else:
|
|
|
|
command += [tb_top_autochecked]
|
|
|
|
run_command("iverilog_verification", "iverilog_output.txt", command)
|
|
|
|
|
2019-09-18 13:14:26 -05:00
|
|
|
vvp_command = ["vvp", compiled_file]
|
2019-08-19 20:06:46 -05:00
|
|
|
output = run_command("vvp_verification", "vvp_sim_output.txt", vvp_command)
|
|
|
|
if "Succeed" in output:
|
|
|
|
logger.info("VVP Simulation Successful")
|
|
|
|
else:
|
2019-09-18 13:14:26 -05:00
|
|
|
logger.error(str(output).split("\n")[-1])
|
|
|
|
if exit_if_fail:
|
|
|
|
clean_up_and_exit("Failed to run VVP verification")
|
2019-08-22 18:01:38 -05:00
|
|
|
ExecTime["VerificationEnd"] = time.time()
|
2019-08-15 15:39:58 -05:00
|
|
|
|
|
|
|
|
|
|
|
def run_command(taskname, logfile, command, exit_if_fail=True):
|
2019-08-21 12:08:13 -05:00
|
|
|
logger.info("Launching %s " % taskname)
|
2019-08-23 00:41:25 -05:00
|
|
|
with open(logfile, 'w+') as output:
|
|
|
|
try:
|
2019-08-15 15:39:58 -05:00
|
|
|
output.write(" ".join(command)+"\n")
|
|
|
|
process = subprocess.run(command,
|
|
|
|
check=True,
|
|
|
|
stdout=subprocess.PIPE,
|
|
|
|
stderr=subprocess.PIPE,
|
|
|
|
universal_newlines=True)
|
|
|
|
output.write(process.stdout)
|
|
|
|
if process.returncode:
|
|
|
|
logger.error("%s run failed with returncode %d" %
|
|
|
|
(taskname, process.returncode))
|
2019-08-23 00:41:25 -05:00
|
|
|
except (Exception, subprocess.CalledProcessError) as e:
|
|
|
|
logger.exception("failed to execute %s" % taskname)
|
|
|
|
filter_failed_process_output(e.output)
|
|
|
|
output.write(e.output)
|
|
|
|
if exit_if_fail:
|
|
|
|
clean_up_and_exit("Failed to run %s task" % taskname)
|
|
|
|
return None
|
2019-08-15 15:39:58 -05:00
|
|
|
logger.info("%s is written in file %s" % (taskname, logfile))
|
2019-08-19 20:06:46 -05:00
|
|
|
return process.stdout
|
2019-08-15 15:39:58 -05:00
|
|
|
|
|
|
|
|
2019-08-23 00:41:25 -05:00
|
|
|
def filter_failed_process_output(vpr_output):
|
2019-08-16 14:36:39 -05:00
|
|
|
for line in vpr_output.split("\n"):
|
|
|
|
if "error" in line.lower():
|
|
|
|
logger.error("-->>" + line)
|
|
|
|
|
2019-08-17 17:22:52 -05:00
|
|
|
|
2019-08-09 17:49:05 -05:00
|
|
|
if __name__ == "__main__":
|
2019-08-22 18:01:38 -05:00
|
|
|
ExecTime["Start"] = time.time()
|
2019-08-15 15:39:58 -05:00
|
|
|
args = parser.parse_args()
|
2019-08-09 17:49:05 -05:00
|
|
|
main()
|