0da9e78e57 | ||
---|---|---|
.. | ||
doc | ||
interfaces | ||
tests | ||
wb_models/housekeepingWB | ||
.gitignore | ||
README.md | ||
caravel.py | ||
caravel_tests.py | ||
caravel_top.sv | ||
cpu.py | ||
pli.tab | ||
sdf_includes.v | ||
tests.json | ||
verify_cocotb.py |
README.md
Overview
Cocotb environment (CTN) is a dynamic simulation testing environment. It's purpose is to speed testing simulation time and get coverage data. The environment is developed using cocotb, an open source coroutine-based co simulation testbench environment for verifying VHDL and SystemVerilog RTL using Python. CTN has 2 main layers: tests and whitebox models. Tests layer contain multiple tests and sequences that can communicate with the caravel (dut) through drivers shown in read at fig 1. Whitebox models layer contain multiple models that should mimic the behavior of each main block inside caravel see fig1. Model is supposed to check if the model is working as expected, if its registers contain the expected values all the time and report coverage of features provided by this block if it’s tested or not.
fig1. caravel testbench environment (read lines are drivers )
Prerequisites
- Docker: Linux || Windows || Mac with Intel Chip || Mac with M1 Chip
- Python 3.6+ with PIP
docker pull efabless/dv:cocotb
- iverilog or vcs
- export CARAVEL_ROOT= <caravel repo root>
- export MCW_ROOT= <caravel_mgmt_soc_litex repo root>
- export PDK_ROOT= <PDK repo root>
- export PDK= <PDK variant sky130A,sky130B,.... >
run a test
Use script verify_cocotb.py
-h, --help show this help message and exit
-regression REGRESSION, -r REGRESSION
name of regression can found in tests.json
-test TEST [TEST ...], -t TEST [TEST ...]
name of test if no --sim provided RTL will be run
<takes list as input>
-sim SIM [SIM ...] Simulation type to be run RTL,GL&GL_SDF provided only
when run -test <takes list as input>
-testlist TESTLIST, -tl TESTLIST
path of testlist to be run
-tag TAG provide tag of the run default would be regression
name and if no regression is provided would be
run_<random float>_<timestamp>_
-maxerr MAXERR max number of errors for every test before simulation
breaks default = 3
-vcs, -v use vcs as compiler if not used iverilog would be used
-cov enable code coverage
-corner CORNER [CORNER ...], -c CORNER [CORNER ...]
Corner type in case of GL_SDF run has to be provided
-keep_pass_unzip Normally the waves and logs of passed tests would be
zipped. Using this option they wouldn't be zipped
Refer to examples
Tests
Refer to tests doc for tests list
cocotb directory tree
├── caravel.py -> contains driving and mentoring functions for caravel interface
├── caravel_top.sv -> testbench top level
├── cpu.py -> contains driving and mentoring functions for wishbone when disable the cpu
├── hex_files -> folder that contains hex files
├── verify_cocotb.py -> script that run tests and regressions
├── sim -> directory get generate when run a test
│ └── <tag> -> tag of the run
│ ├── <sim type>-<test name> -> test result directory contain all logs and wave related to the test
│ ├── command.log -> command use for this run
│ └── runs.log -> contains status of the run fails and passes tests
├── tests -> directory contains all the tests
├── tests.json -> test list have all the tests, regressions and contain small description about every test
└── wb_models -> contains checkers and models for some caravel blocks
How to debug
TO BE ADDED