Deep learning toolkit-enabled VLSI placement.
With the analogy between nonlinear VLSI placement and deep learning training problem, this tool is developed with deep learning toolkit for flexibility and efficiency.
The tool runs on both CPU and GPU.
Over 30X speedup over the CPU implementation (RePlAce) is achieved in global placement and legalization on ISPD 2005 contest benchmarks with a Nvidia Tesla V100 GPU.
DREAMPlace runs on both CPU and GPU. If it is installed on a machine without GPU, only CPU support will be enabled with multi-threading.
| Bigblue4 | Density Map | Electric Potential | Electric Field |
|---|---|---|---|
-
Yibo Lin, Shounak Dhar, Wuxi Li, Haoxing Ren, Brucek Khailany and David Z. Pan, "DREAMPlace: Deep Learning Toolkit-Enabled GPU Acceleration for Modern VLSI Placement", ACM/IEEE Design Automation Conference (DAC), Las Vegas, NV, Jun 2-6, 2019 (preprint) (slides)
-
Yibo Lin, Zixuan Jiang, Jiaqi Gu, Wuxi Li, Shounak Dhar, Haoxing Ren, Brucek Khailany and David Z. Pan, "DREAMPlace: Deep Learning Toolkit-Enabled GPU Acceleration for Modern VLSI Placement", IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems (TCAD), 2020 (in submission)
-
Python 2.7 or Python 3.5/3.6/3.7
-
Pytorch 1.0.0
- Other version around 1.0.0 may also work, but not tested
-
- Recommend GCC 5.1 or later.
- Other compilers may also work, but not tested.
-
- Need to install and visible for linking
-
- Integrated as a git submodule
-
- Integrated as a submodule
-
CUDA 9.1 or later (Optional)
- If installed and found, GPU acceleration will be enabled.
- Otherwise, only CPU implementation is enabled.
-
GPU architecture compatibility 6.0 or later (Optional)
- Code has been tested on GPUs with compute compatibility 6.0 and 7.0.
- Please check the compatibility of the GPU devices.
- The default compilation target is compatibility 6.0. For compatibility 7.0, it is necessary to set the CMAKE_CUDA_FLAGS to -gencode=arch=compute_70,code=sm_70.
-
Cairo (Optional)
- If installed and found, the plotting functions will be faster by using C/C++ implementation.
- Otherwise, python implementation is used.
-
NTUPlace3 (Optional)
- If the binary is provided, it can be used to perform detailed placement.
To pull git submodules in the root directory
git submodule init
git submodule update
Or alternatively, pull all the submodules when cloning the repository.
git clone --recursive https://github.com/limbo018/DREAMPlace.git
Go to the root directory.
pip install -r requirements.txt
Two options are provided for building: with and without Docker.
You can use the Docker container to avoid building all the dependencies yourself.
- Install Docker on Windows, Mac or Linux.
- To enable the GPU features, install NVIDIA-docker; otherwise, skip this step.
- Navigate to the repository.
- Get the docker container with either of the following options.
- Option 1: pull from the cloud limbo018/dreamplace.
docker pull limbo018/dreamplace:cuda- Option 2: build the container.
docker build . --file Dockerfile --tag your_name/dreamplace:cuda - Enter bash environment of the container. Replace
limbo018with your name if option 2 is chosen in the previous step.
Run with GPU.
docker run --gpus 1 -it -v $(pwd):/DREAMPlace limbo018/dreamplace:cuda bash
Run without GPU.
docker run -it -v $(pwd):/DREAMPlace limbo018/dreamplace:cuda bash
cd /DREAMPlace.- Go to next section to complete building.
CMake is adopted as the makefile system. To build, go to the root directory.
mkdir build
cd build
cmake .. -DCMAKE_INSTALL_PREFIX=your_install_path
make
make install
Third party submodules are automatically built except for Boost.
To clean, go to the root directory.
rm -r build
Here are the available options for CMake.
- CMAKE_INSTALL_PREFIX: installation directory
- Example
cmake -DCMAKE_INSTALL_PREFIX=path/to/your/directory
- Example
- CMAKE_CUDA_FLAGS: custom string for NVCC (default -gencode=arch=compute_60,code=sm_60)
- Example
cmake -DCMAKE_CUDA_FLAGS=-gencode=arch=compute_60,code=sm_60
- Example
- CMAKE_CXX_ABI: 0|1 for the value of _GLIBCXX_USE_CXX11_ABI for C++ compiler, default is 0.
- Example
cmake -DCMAKE_CXX_ABI=0 - It must be consistent with the _GLIBCXX_USE_CXX11_ABI for compling all the C++ dependencies, such as Boost and PyTorch.
- PyTorch in default is compiled with _GLIBCXX_USE_CXX11_ABI=0, but in a customized PyTorch environment, it might be compiled with _GLIBCXX_USE_CXX11_ABI=1.
- Example
To get ISPD 2005 benchmarks, run the following script from the directory.
python benchmarks/ispd2005.py
Before running, make sure the benchmarks have been downloaded and the python dependency packages have been installed. Go to the install directory and run with JSON configuration file for full placement.
python dreamplace/Placer.py test/ispd2005/adaptec1.json
Test individual pytorch op with the unitest in the root directory.
python unitest/ops/hpwl_unitest.py
Descriptions of options in JSON configuration file can be found by running the following command.
python dreamplace/Placer.py --help
- Yibo Lin, supervised by David Z. Pan, composed the initial release.
- Zixuan Jiang and Jiaqi Gu improved the efficiency of the wirelength and density operators on GPU.
- Pull requests to improve the tool are more than welcome. We appreciate all kinds of contributions from the community.
-
- Multi-threaded CPU and optional GPU acceleration support
-
- Net weighting support through .wts files in Bookshelf format
- Incremental placement support
-
- LEF/DEF support as input/output
- Python binding and access to C++ placement database
-
- Improved efficiency for wirelength and density operators from TCAD extension
-
- Docker container for building environment