blob: 497a90895def4c5bab9b19a18acc3d0e38d8090d [file] [log] [blame]
#!/usr/bin/env python3
#
# foundry_install.py
#
# This file generates the local directory structure and populates the
# directories with foundry vendor data. The local directory (target)
# should be a staging area, not a place where files are kept permanently.
#
# Options:
# -ef_format Use efabless naming (libs.ref/techLEF),
# otherwise use generic naming (libs.tech/lef)
# -timestamp <value> Pass a timestamp to use for stamping GDS and MAG files
# -clean Clear out and remove target directory before starting
# -source <path> Path to source data top level directory
# -target <path> Path to target (staging) top level directory
#
# All other options represent paths to vendor files. They may all be
# wildcarded with "*", or with specific escapes like "%l" for library
# name or "%v" for version number (see below for a complete list of escape
# sequences).
#
# Note only one of "-spice" or "-cdl" need be specified. Since the
# open source tools use ngspice, CDL files are converted to ngspice
# syntax when needed.
#
# -techlef <path> Path to technology LEF file
# -doc <path> Path to technology documentation
# -lef <path> Path to LEF file
# -spice <path> Path to SPICE netlists
# -cdl <path> Path to CDL netlists
# -models <path> Path to SPICE (primitive device) models
# -liberty <path> Path to Liberty timing files
# -gds <path> Path to GDS data
# -verilog <path> Path to verilog models
#
# -library <type> <name> [<target>] See below
#
# For the "-library" option, any number of libraries may be supported, and
# one "-library" option should be provided for each supported library.
# <type> is one of: "digital", "primitive", or "general". Analog and I/O
# libraries fall under the category "general", as they are all treated the
# same way. <name> is the vendor name of the library. [<target>] is the
# (optional) local name of the library. If omitted, then the vendor name
# is used for the target (there is no particular reason to specify a
# different local name for a library).
#
# In special cases using options (see below), path may be "-", indicating
# that there are no source files, but only to run compilations or conversions
# on the files in the target directory.
#
# All options "-lef", "-spice", etc., can take the additional arguments
# up=<number>
#
# to indicate that the source hierarchy should be copied from <number>
# levels above the files. For example, if liberty files are kept in
# multiple directories according to voltage level, then
#
# -liberty x/y/z/PVT_*/*.lib
#
# would install all .lib files directly into libs.ref/<libname>/liberty/*.lib
# (if "-ef_format" option specified, then: libs.ref/<libname>/liberty/*.lib)
# while
#
# -liberty x/y/z/PVT_*/*.lib up=1
#
# would install all .lib files into libs.ref/liberty/<libname>/PVT_*/*.lib
# (if "-ef_format" option specified, then: libs.ref/<libname>/liberty/PVT_*/*.lib)
#
# Please note that the INSTALL variable in the Makefile starts with "set -f"
# to suppress the OS from doing wildcard substitution; otherwise the
# wildcards in the install options will get expanded by the OS before
# being passed to the install script.
#
# Other library-specific arguments are:
#
# nospec : Remove timing specification before installing
# (used with verilog files; needs to be extended to
# liberty files)
#
# compile : Create a single library from all components. Used
# when a foundry library has inconveniently split
# an IP library (LEF, CDL, verilog, etc.) into
# individual files.
#
# compile-only: Like "compile" except that the individual
# files are removed after the library file has been
# created (NOTE: equivalent to "compile remove").
#
# stub : Remove contents of subcircuits from CDL or SPICE
# netlist files.
#
# priv : Mark the contents being installed as privleged, and
# put them in a separate root directory libs.priv
# where they can be given additional read/write
# restrictions.
#
# exclude : Followed by "=" and a comma-separated list of names.
# exclude these files/modules/subcircuits. Names may
# also be wildcarded in "glob" format. Cell names are
# excluded from the copy and also from any libraries
# that are generated. If the pattern contains a directory
# path, then patterns are added from files in the
# specified directory instead of the source.
#
# no-copy : Followed by "=" and a comma-separated list of names.
# exclude these files/modules/subcircuits. Names may
# also be wildcarded in "glob" format. Cell names are
# excluded from the copy only. If the pattern contains
# a directory path, then patterns are added from files
# in the specified directory instead of the source.
#
# include : Followed by "=" and a comma-separated list of names.
# include these files/modules/subcircuits. Names may
# also be wildcarded in "glob" format. Cell names are
# included in any libraries that are generated if they
# exist in the target directory, even if they were not
# in the source directory being copied. If the pattern
# contains a directory path, then patterns are added from
# files in the specified directory instead of the source.
#
# header : Followed by "=" and an alternative name. When compiling
# a library file, use this file as a header. Required for
# liberty format; optional for others. Behavior is format-
# dependent; e.g., for liberty headers, the cell entries
# will be inserted before the final closing brace. The name
# may be wildcarded, but if so, there must only be a single
# matching file.
#
# rename : Followed by "=" and an alternative name. For any
# file that is a single entry, change the name of
# the file in the target directory to this. If the
# alternative name is a file extension (optionally
# preceded by '*', as in "rename=*.spice", then all
# input files are renamed with the indicated file
# extension (To-do: take regexps for multiple files).
# When used with "compile" or "compile-only", this
# refers to the name of the target compiled file.
#
# filter: Followed by "=" and the name of a script.
# Each file is passed through the filter script
# before writing into the staging area.
#
# sort: Followed by "=" and the name of a script.
# The list of files to process (after applying items
# from "exclude") will be written to a file
# "filelist.txt", which will be used by the
# library compile routines, if present. The sort
# script will rewrite the file with the order in
# which entries should appear in the compiled library.
# Only useful when used with "compile" or "compile-only".
# If not specified, files are sorted by "natural sort"
# order.
#
# annotate: When used with "-lef", the LEF files are used only
# for annotation of pins (use, direction, etc.), but
# the LEF files should not be used and LEF should be
# generated from layout.
#
# lefopts: Followed by "=" and a comma-separated list of option
# strings. If LEF views are generated from magic, use the
# options specified.
#
# noconvert: Install only; do not attempt to convert to other
# formats (applies only to GDS, CDL, and LEF).
#
# dorcx: Used with "-gds" and when no CDL or SPICE files are
# specified for installation: Do parasitic extraction
# (NOTE: Only does parasitic capacitance extraction).
#
# noextract: Used with "-gds". Do not extract the layout into a
# netlist (implies that the layout uses the "device
# primitive" property).
#
# options: Followed by "=" and the name of a script. Behavior
# is dependent on the mode; if applied to "-gds",
# then the script is inserted before the GDS read
# in the Tcl generate script passed to magic. If
# what follows the "=" is not a file, then it is
# Tcl code to be inserted verbatim.
#
# collate: Files in the source may be split in parts and such
# files should be be collated before copying.
#
# remove: Indicates that files are copied to the staging area
# and removed after all operations. This allows files
# that are excluded from a library but are handled by
# a filter script to be cleaned up afterward.
#
# NOTE: This script can be called once for all libraries if all file
# types (gds, cdl, lef, etc.) happen to all work with the same wildcards.
# However, it is more likely that it will be called several times for the
# same PDK, once to install I/O cells, once to install digital, and so
# forth, as made possible by the wild-carding.
import re
import os
import sys
import glob
import stat
import shutil
import fnmatch
import subprocess
# Import local routines
import natural_sort
from create_gds_library import create_gds_library
from create_spice_library import create_spice_library
from create_lef_library import create_lef_library
from create_lib_library import create_lib_library
from create_verilog_library import create_verilog_library
def usage():
print("foundry_install.py [options...]")
print(" -copy Copy files from source to target (default)")
print(" -ef_format Use efabless naming conventions for local directories")
print(" -timestamp <value> Use <value> for timestamping files")
print("")
print(" -source <path> Path to top of source directory tree")
print(" -target <path> Path to top of target directory tree")
print("")
print(" -techlef <path> Path to technology LEF file")
print(" -doc <path> Path to technology documentation")
print(" -lef <path> Path to LEF file")
print(" -spice <path> Path to SPICE netlists")
print(" -cdl <path> Path to CDL netlists")
print(" -models <path> Path to SPICE (primitive device) models")
print(" -lib <path> Path to Liberty timing files")
print(" -liberty <path> Path to Liberty timing files")
print(" -gds <path> Path to GDS data")
print(" -verilog <path> Path to verilog models")
print(" -library <type> <name> [<target>] See below")
print("")
print(" All <path> names may be wild-carded with '*' ('glob'-style wild-cards)")
print("")
print(" All options with <path> other than source and target may take the additional")
print(" arguments 'up=<number>', where <number> indicates the number of levels of")
print(" hierarchy of the source path to include when copying to the target.")
print("")
print(" Library <type> may be one of:")
print(" digital Digital standard cell library")
print(" primitive Primitive device library")
print(" general All other library types (I/O, analog, etc.)")
print("")
print(" If <target> is unspecified then <name> is used for the target.")
# Return a list of files after glob-style substituting into pathname. This
# mostly relies on glob.glob(), but uses the additional substitutions with
# escape strings:
#
# %v : Match a version number in the form "major[.minor[.rev]]"
# %l : substitute the library name
# %% : substitute the percent character verbatim
try:
from setuptools.distutils.version import LooseVersion
except:
from distutils.version import LooseVersion
#----------------------------------------------------------------------------
#----------------------------------------------------------------------------
def subprocess_run(name, cmd, stdin=subprocess.DEVNULL, cwd=None):
# Make sure the output pipes are flushed before running the command to make
# output look neater / more ordered.
sys.stdout.flush()
sys.stderr.flush()
fproc = subprocess.Popen(
cmd,
stdin = stdin,
stdout = subprocess.PIPE,
stderr = subprocess.PIPE,
universal_newlines = True,
cwd = cwd or os.curdir,
)
stdout, stderr = fproc.communicate()
if stdout:
for line in stdout.splitlines():
print(line)
if stderr:
print('Error message output from {} script:'.format(name))
for line in stderr.splitlines():
print(line)
if fproc.returncode != 0:
emsg = [
"Command {} failed with exit code: {}\n".format(
name, fproc.returncode),
" " + " ".join(cmd),
]
if stdin != subprocess.DEVNULL:
stdin.seek(0)
input_script = stdin.read()
emsg += [
"\nInput script was:\n",
'-'*75,'\n',
input_script,'\n',
'-'*75,'\n',
]
raise SystemError("".join(emsg))
#----------------------------------------------------------------------------
#----------------------------------------------------------------------------
def makeuserwritable(filepath):
if os.path.exists(filepath):
st = os.stat(filepath)
os.chmod(filepath, st.st_mode | stat.S_IWUSR)
#----------------------------------------------------------------------------
#----------------------------------------------------------------------------
def substitute(pathname, library):
if library:
# Do %l substitution
newpathname = re.sub('%l', library, pathname)
else:
newpathname = pathname
if '%v' in newpathname:
vglob = re.sub('%v.*', '*', newpathname)
vlibs = glob.glob(vglob)
try:
vstr = vlibs[0][len(vglob)-1:]
except IndexError:
pass
else:
for vlib in vlibs[1:]:
vtest = vlib[len(vglob)-1:]
if LooseVersion(vtest) > LooseVersion(vstr):
vstr = vtest
newpathname = re.sub('%v', vstr, newpathname)
if '%%' in newpathname:
newpathname = re.sub('%%', '%', newpathname)
return newpathname
#----------------------------------------------------------------------------
#----------------------------------------------------------------------------
def get_gds_properties(magfile):
proprex = re.compile('^[ \t]*string[ \t]+(GDS_[^ \t]+)[ \t]+([^ \t]+)$')
proplines = []
if os.path.isfile(magfile):
with open(magfile, 'r') as ifile:
magtext = ifile.read().splitlines()
for line in magtext:
lmatch = proprex.match(line)
if lmatch:
propline = lmatch.group(1) + ' ' + lmatch.group(2)
proplines.append(propline)
return proplines
#----------------------------------------------------------------------------
# Read subcircuit ports from a CDL file, given a subcircuit name that should
# appear in the file as a subcircuit entry, and return a dictionary of ports
# and their indexes in the subcircuit line.
#----------------------------------------------------------------------------
def get_subckt_ports(cdlfile, subname):
portdict = {}
pidx = 1
portrex = re.compile('^\.subckt[ \t]+([^ \t]+)[ \t]+(.*)$', flags=re.IGNORECASE)
with open(cdlfile, 'r') as ifile:
cdltext = ifile.read()
cdllines = cdltext.replace('\n+', ' ').splitlines()
for line in cdllines:
lmatch = portrex.match(line)
if lmatch:
if lmatch.group(1).lower() == subname.lower():
ports = lmatch.group(2).split()
for port in ports:
portdict[port.lower()] = pidx
pidx += 1
break
return portdict
#----------------------------------------------------------------------------
# Filter a verilog file to remove any backslash continuation lines, which
# iverilog does not parse. If targetroot is a directory, then find and
# process all files in the path of targetroot. If any file to be processed
# is unmodified (has no backslash continuation lines), then ignore it. If
# any file is a symbolic link and gets modified, then remove the symbolic
# link before overwriting with the modified file.
#----------------------------------------------------------------------------
def vfilefilter(vfile):
modified = False
with open(vfile, 'r') as ifile:
vtext = ifile.read()
# Remove backslash-followed-by-newline and absorb initial whitespace. It
# is unclear what initial whitespace means in this context, as the use-
# case that has been seen seems to work under the assumption that leading
# whitespace is ignored up to the amount used by the last indentation.
vlines = re.sub('\\\\\n[ \t]*', '', vtext)
if vlines != vtext:
# File contents have been modified, so if this file was a symbolic
# link, then remove it. Otherwise, overwrite the file with the
# modified contents.
makeuserwritable(vfile)
if os.path.islink(vfile):
os.unlink(vfile)
with open(vfile, 'w') as ofile:
ofile.write(vlines)
#----------------------------------------------------------------------------
# Run a filter on verilog files that cleans up known syntax issues.
# This is embedded in the foundry_install script and is not a custom
# filter largely because the issue is in the tool, not the PDK.
#----------------------------------------------------------------------------
def vfilter(targetroot):
if os.path.isfile(targetroot):
vfilefilter(targetroot)
else:
vlist = glob.glob(targetroot + '/*')
for vfile in vlist:
if os.path.isfile(vfile):
vfilefilter(vfile)
#----------------------------------------------------------------------------
# For issues that are PDK-specific, a script can be written and put in
# the PDK's custom/scripts/ directory, and passed to the foundry_install
# script using the "filter" option.
#----------------------------------------------------------------------------
def tfilter(targetroot, filterscript, ef_format=False, outfile=[]):
filterroot = os.path.split(filterscript)[1]
if os.path.isfile(targetroot):
print(' Filtering file ' + targetroot + ' with ' + filterroot)
if not outfile:
outfile = targetroot
else:
# Make sure this file is writable (as the original may not be)
makeuserwritable(outfile)
arguments = [filterscript, targetroot, outfile]
if ef_format:
arguments.append('-ef_format')
subprocess_run('filter', arguments)
else:
tlist = glob.glob(targetroot + '/*')
for tfile in tlist:
if os.path.isfile(tfile):
print(' Filtering file ' + tfile + ' with ' + filterroot)
subprocess_run('filter', [filterscript, tfile, tfile])
#----------------------------------------------------------------------------
# This is the main entry point for the foundry install script.
#----------------------------------------------------------------------------
if __name__ == '__main__':
if len(sys.argv) == 1:
print("No options given to foundry_install.py.")
usage()
sys.exit(0)
optionlist = []
newopt = []
removelist = []
sourcedir = None
targetdir = None
ef_format = False
do_timestamp = False
timestamp_value = 0
do_clean = False
lef_savelibname = None
have_lef = False
have_techlef = False
have_lefanno = False
have_gds = False
have_spice = False
have_cdl = False
have_verilog = False
have_lib = False
# Break arguments into groups where the first word begins with "-".
# All following words not beginning with "-" are appended to the
# same list (optionlist). Then each optionlist is processed.
# Note that the first entry in optionlist has the '-' removed.
for option in sys.argv[1:]:
if option.find('-', 0) == 0:
if newopt != []:
optionlist.append(newopt)
newopt = []
newopt.append(option[1:])
else:
newopt.append(option)
if newopt != []:
optionlist.append(newopt)
# Pull library names from optionlist
libraries = []
for option in optionlist[:]:
if option[0] == 'library':
optionlist.remove(option)
libraries.append(option[1:])
# Check for option "ef_format" or "std_format" or "clean"
for option in optionlist[:]:
if option[0] == 'ef_naming' or option[0] == 'ef_names' or option[0] == 'ef_format':
optionlist.remove(option)
ef_format = True
elif option[0] == 'std_naming' or option[0] == 'std_names' or option[0] == 'std_format':
optionlist.remove(option)
ef_format = False
elif option[0] == 'timestamp':
optionlist.remove(option)
if len(option) > 1:
timestamp_value = option[1]
do_timestamp = True
else:
print('Error: Option "timestamp" used with no value.')
elif option[0] == 'clean':
do_clean = True
# Check for options "source" and "target"
for option in optionlist[:]:
if option[0] == 'source':
optionlist.remove(option)
if len(option) > 1:
sourcedir = option[1]
else:
print('Error: Option "source" used with no value.')
elif option[0] == 'target':
optionlist.remove(option)
if len(option) > 1:
targetdir = option[1]
else:
print('Error: Option "target" used with no value.')
if not targetdir:
print("No target directory specified. Exiting.")
sys.exit(1)
# Take the target PDK name from the target path last component
pdkname = os.path.split(targetdir)[1]
# If targetdir (the staging area) exists, make sure it's empty.
if os.path.isdir(targetdir):
# Error if targetdir exists but is not writeable
if not os.access(targetdir, os.W_OK):
print("Target installation directory " + targetdir + " is not writable.")
sys.exit(1)
# Clear out the staging directory if specified
if do_clean:
shutil.rmtree(targetdir)
elif os.path.exists(targetdir):
print("Target installation directory " + targetdir + " is not a directory.")
sys.exit(1)
# Error if no source or dest specified unless "-clean" was specified
if not sourcedir:
if do_clean:
print("Done removing staging area.")
sys.exit(0)
else:
print("No source directory specified. Exiting.")
sys.exit(1)
# Create the target directory
os.makedirs(targetdir, exist_ok=True)
# Here's where common scripts are found:
openpdksdir = os.path.dirname(os.path.realpath(__file__))
scriptdir = os.path.split(openpdksdir)[0] + '/common'
#----------------------------------------------------------------
# Installation part 1: Install files into the staging directory
#----------------------------------------------------------------
# Diagnostic
print("Installing in target (staging) directory " + targetdir)
# Create the top-level directories
os.makedirs(targetdir + '/libs.tech', exist_ok=True)
os.makedirs(targetdir + '/libs.ref', exist_ok=True)
# Path to magic techfile depends on ef_format
if ef_format == True:
mag_current = '/libs.tech/magic/current/'
else:
mag_current = '/libs.tech/magic/'
# Check for magic version and set flag if it does not exist or if
# it has the wrong version.
have_mag_8_2 = False
have_mag_8_3_261 = False
try:
mproc = subprocess.run(
['magic', '--version'],
stdout = subprocess.PIPE,
stderr = subprocess.PIPE,
universal_newlines = True,
check = True,
)
if mproc.stdout:
mag_version = mproc.stdout.splitlines()[0]
mag_version_info = mag_version.split('.')
try:
if int(mag_version_info[0]) > 8:
have_mag_8_2 = True
elif int(mag_version_info[0]) == 8:
if int(mag_version_info[1]) >= 2:
have_mag_8_2 = True
print('Magic version 8.2 (or better) available on the system.')
if int(mag_version_info[0]) > 8:
have_mag_8_3_261 = True
elif int(mag_version_info[0]) == 8:
if int(mag_version_info[1]) > 3:
have_mag_8_3_261 = True
elif int(mag_version_info[1]) == 3:
if int(mag_version_info[2]) >= 261:
have_mag_8_3_261 = True
except ValueError:
print('Error: "magic --version" did not return valid version number.')
if mproc.stderr:
print(mproc.stderr)
except FileNotFoundError:
print('Error: Failed to find executable for magic in standard search path.')
if not have_mag_8_2:
print('WARNING: Magic version 8.2 (or beter) cannot be executed ')
print('from the standard executable search path.')
print('Please install or correct the search path.')
print('Magic database files will not be created, and other missing file formats may not be generated.')
# Populate any targets that do not specify a library, or where the library is
# specified as "primitive".
# Populate the techLEF and SPICE models, if specified. Also, this section can add
# to any directory in libs.tech/ as given by the option; e.g., "-ngspice" will
# install into libs.tech/ngspice/.
if libraries == [] or 'primitive' in libraries[0]:
for option in optionlist[:]:
# Legacy behavior is to put libs.tech models and techLEF files in
# the same grouping as files for the primdev library (which go in
# libs.ref). Current behavior is to put all libs.tech files in
# a grouping with no library, with unrestricted ability to write
# into any subdirectory of libs.tech/. Therefore, need to restrict
# legacy use to just 'techlef' and 'models'.
if len(libraries) > 0 and 'primitive' in libraries[0]:
if option[0] != 'techlef' and option[0] != 'techLEF' and option[0] != 'models':
continue
# Normally technology LEF files are associated with IP libraries.
# However, if no library is specified or the library is 'primitive'
# (legacy behavior), then put in the techLEF directory with no subdirectory.
filter_scripts = []
if option[0] == 'techlef' or option[0] == 'techLEF':
for item in option:
if item.split('=')[0] == 'filter':
filter_scripts.append(item.split('=')[1])
if ef_format:
techlefdir = targetdir + '/libs.ref/' + 'techLEF'
else:
techlefdir = targetdir + '/libs.tech/lef'
os.makedirs(techlefdir, exist_ok=True)
# All techlef files should be copied, so use "glob" on the wildcards
techlist = glob.glob(substitute(sourcedir + '/' + option[1], None))
for lefname in techlist:
leffile = os.path.split(lefname)[1]
targname = techlefdir + '/' + leffile
if os.path.isfile(lefname):
shutil.copy(lefname, targname)
else:
shutil.copytree(lefname, targname)
for filter_script in filter_scripts:
# Apply filter script to all files in the target directory
tfilter(targname, filter_script, ef_format)
optionlist.remove(option)
# All remaining options will refer to specific tools (e.g., -ngspice, -magic)
# although generic names (.e.g, -models) are acceptable if the tools know
# where to find the files. Currently, most tools have their own formats
# and standards for setup, and so generally each install directory will be
# unique to one EDA tool.
else:
filter_scripts = []
for item in option:
if item.split('=')[0] == 'filter':
filter_scripts.append(item.split('=')[1])
print('Diagnostic: installing to ' + option[0] + '.')
tooldir = targetdir + '/libs.tech/' + option[0]
os.makedirs(tooldir, exist_ok=True)
# All files should be linked or copied, so use "glob" on
# the wildcards. Copy each file and recursively copy each
# directory.
toollist = glob.glob(substitute(sourcedir + '/' + option[1], None))
for toolname in toollist:
toolfile = os.path.split(toolname)[1]
targname = tooldir + '/' + toolfile
print(' installing from ' + toolfile + ' to ' + targname)
if os.path.isdir(toolname):
# Remove any existing directory, and its contents
if os.path.isdir(targname):
shutil.rmtree(targname)
os.makedirs(targname)
# Recursively find and copy or link the whole directory
# tree from this point.
alltoollist = glob.glob(toolname + '/**', recursive=True)
commonpart = os.path.commonpath(alltoollist)
for subtoolname in alltoollist:
# Get the path part that is not common between toollist and
# alltoollist.
subpart = os.path.relpath(subtoolname, commonpart)
subtargname = targname + '/' + subpart
if os.path.isfile(subtoolname):
os.makedirs(os.path.split(subtargname)[0], exist_ok=True)
shutil.copy(subtoolname, subtargname)
else:
print(' copy tree from ' + subtoolname + ' to ' + subtargname)
# emulate Python3.8 dirs_exist_ok option
try:
shutil.copytree(subtoolname, subtargname)
except FileExistsError:
pass
for filter_script in filter_scripts:
# Apply filter script to all files in the target directory
tfilter(subtargname, filter_script, ef_format)
else:
# Remove any existing file
if os.path.isfile(targname):
os.remove(targname)
elif os.path.isdir(targname):
shutil.rmtree(targname)
if os.path.isfile(toolname):
shutil.copy(toolname, targname)
else:
shutil.copytree(toolname, targname)
for filter_script in filter_scripts:
# Apply filter script to all files in the target directory
tfilter(targname, filter_script, ef_format)
optionlist.remove(option)
# Do an initial pass through all of the options and determine what is being
# installed, so that we know in advance which file formats are missing and
# need to be generated.
for option in optionlist[:]:
if option[0] == 'lef':
have_lefanno = True if 'annotate' in option or 'anno' in option else False
have_lef = True if not have_lefanno else False
if option[0] == 'techlef' or option[0] == 'techLEF':
have_techlef = True
elif option[0] == 'gds':
have_gds = True
elif option[0] == 'spice' or option[0] == 'spi':
have_spice = True
elif option[0] == 'cdl':
have_cdl = True
elif option[0] == 'verilog':
have_verilog = True
elif option[0] == 'lib' or option[0] == 'liberty':
have_lib = True
# The remaining options in optionlist should all be types like 'lef' or 'liberty'
# and there should be a corresponding library list specified by '-library'
for option in optionlist[:]:
# Ignore if no library list---should have been taken care of above.
if libraries == []:
break
# Diagnostic
print("Install option: " + str(option[0]))
if option[0] == 'lef' and have_lefanno:
print("LEF files used for annotation only. Temporary install.")
# For ef_format: always make techlef -> techLEF and spice -> spi
if ef_format:
if option[0] == 'techlef':
option[0] = 'techLEF'
elif option[0] == 'spice':
option[0] = 'spi'
destdir = targetdir + '/libs.ref/' + option[0]
os.makedirs(destdir, exist_ok=True)
# If the option is followed by the keyword "up" and a number, then
# the source should be copied (or linked) from <number> levels up
# in the hierarchy (see below).
hier_up = 0
for item in option:
if item.split('=')[0] == 'up':
hier_up = int(item.split('=')[1])
break
filter_scripts = []
for item in option:
if item.split('=')[0] == 'filter':
filter_scripts.append(item.split('=')[1])
# Option 'collate' is a standalone keyword
do_collate = 'collate' in option
# Option 'stub' applies to netlists ('cdl' or 'spice') and generates
# a file with only stub entries.
do_stub = 'stub' in option
# Option 'compile' is a standalone keyword ('comp' may be used).
do_compile = 'compile' in option or 'comp' in option
do_compile_only = 'compile-only' in option or 'comp-only' in option
# Option 'nospecify' is a standalone keyword ('nospec' may be used).
do_remove_spec = 'nospecify' in option or 'nospec' in option
# Option 'remove' is a standalone keyword.
do_remove = 'remove' in option
# Option 'exclude' has an argument
excludelist = []
try:
excludelist = list(item.split('=')[1].split(',') for item in option if item.startswith('excl'))[0]
except IndexError:
pass
if len(excludelist) > 0:
print('Excluding files: ' + (',').join(excludelist))
# Option 'no-copy' has an argument
nocopylist = []
try:
nocopylist = list(item.split('=')[1].split(',') for item in option if item.startswith('no-copy'))[0]
except IndexError:
pass
if len(nocopylist) > 0:
print('Not copying files: ' + (',').join(nocopylist))
# Option 'include' has an argument
includelist = []
try:
includelist = list(item.split('=')[1].split(',') for item in option if item.startswith('incl'))[0]
except IndexError:
pass
if len(includelist) > 0:
print('Including files: ' + (',').join(includelist))
# Option 'header' has an argument
headerfile = None
try:
headerfile = list(item.split('=')[1] for item in option if item.startswith('header'))[0]
except IndexError:
headerfile = None
else:
print('Header file is: ' + headerfile)
# Option 'rename' has an argument
try:
newname = list(item.split('=')[1] for item in option if item.startswith('rename'))[0]
except IndexError:
newname = None
else:
print('Renaming file to: ' + newname)
# Option 'sort' has an argument. . .
try:
sortscript = list(item.split('=')[1] for item in option if item.startswith('sort'))[0]
except IndexError:
# If option 'sort' is not specified, then use the "natural sort" script
sortscript = scriptdir + '/sort_pdkfiles.py'
else:
print('Sorting files with script ' + sortscript)
# For each library, create the library subdirectory
for library in libraries:
if len(library) == 3:
destlib = library[2]
else:
destlib = library[1]
if ef_format:
destlibdir = destdir + '/' + destlib
else:
destdir = targetdir + '/libs.ref/' + destlib + '/' + option[0]
destlibdir = destdir
os.makedirs(destlibdir, exist_ok=True)
# Populate the library subdirectory
# Parse the option and replace each '/*/' with the library name,
# and check if it is a valid directory name. Then glob the
# resulting option name. Warning: This assumes that all
# occurences of the text '/*/' match a library name. It should
# be possible to wild-card the directory name in such a way that
# this is always true.
testpath = substitute(sourcedir + '/' + option[1], library[1])
liblist = glob.glob(testpath)
liblist = natural_sort.natural_sort(liblist)
if do_collate:
# Rework liblist so that file parts are put into a list and
# associated with the final filename. Replace the file parts
# in liblist with the collated filename. Assume that the
# parts are named with an additional extension such as ".part1"
# ".part2" etc., and that these extensions are in natural sort
# order. Note that it is not necessary for all files in the
# list to be split into parts.
# Regular expression catches filename "a.b.c", handling any
# leading dots (such as "../") and returns string "a.b"
basenamerex = re.compile('(\.?\.?[^.]+\..+)\.+')
baseliblist = []
splitfiles = {}
for libfile in liblist:
bmatch = basenamerex.match(libfile)
if bmatch:
basename = bmatch.group(1)
if basename not in baseliblist:
baseliblist.append(basename)
splitfiles[basename] = [libfile]
else:
splitfiles[basename].append(libfile)
else:
baseliblist.append(libfile) ;# not a split file
liblist = baseliblist
# Create a file "sources.txt" (or append to it if it exists)
# and add the source directory name so that the staging install
# script can know where the files came from.
with open(destlibdir + '/sources.txt', 'a') as ofile:
print(testpath, file=ofile)
# Create exclude list with glob-style matching using fnmatch
if len(liblist) > 0:
liblistnames = list(os.path.split(item)[1] for item in liblist)
notliblist = []
for exclude in excludelist:
if '/' in exclude:
# Names come from files in a path that is not the source
excludefiles = os.listdir(os.path.split(exclude)[0])
pattern = os.path.split(exclude)[1]
notliblist.extend(fnmatch.filter(excludefiles, pattern))
else:
notliblist.extend(fnmatch.filter(liblistnames, exclude))
# Apply exclude list
if len(notliblist) > 0:
for file in liblist[:]:
if os.path.split(file)[1] in notliblist:
liblist.remove(file)
if len(excludelist) > 0 and len(notliblist) == 0:
print('Warning: Nothing from the exclude list found in sources.')
print('excludelist = ' + str(excludelist))
print('destlibdir = ' + destlibdir)
# Create a list of cell names not to be copied from "nocopylist"
nocopynames = []
for nocopy in nocopylist:
if '/' in nocopy:
# Names come from files in a path that is not the source
nocopyfiles = os.listdir(os.path.split(nocopy)[0])
pattern = os.path.split(nocopy)[1]
nocopynames.extend(fnmatch.filter(nocopyfiles, pattern))
else:
nocopynames.extend(fnmatch.filter(liblistnames, nocopy))
# Diagnostic
print('Collecting files from ' + testpath)
print('Files to install:')
if len(liblist) < 10:
for item in liblist:
print(' ' + item)
else:
for item in liblist[0:4]:
print(' ' + item)
print(' .')
print(' .')
print(' .')
for item in liblist[-6:-1]:
print(' ' + item)
print('(' + str(len(liblist)) + ' files total)')
destfilelist = []
for libname in liblist:
# Note that there may be a hierarchy to the files in option[1],
# say for liberty timing files under different conditions, so
# make sure directories have been created as needed.
libfile = os.path.split(libname)[1]
libfilepath = os.path.split(libname)[0]
destpathcomp = []
for i in range(hier_up):
destpathcomp.append('/' + os.path.split(libfilepath)[1])
libfilepath = os.path.split(libfilepath)[0]
destpathcomp.reverse()
destpath = ''.join(destpathcomp)
dontcopy = True if libfile in nocopynames else False
if option[0] == 'verilog':
fileext = '.v'
elif option[0] == 'liberty' or option[0] == 'lib':
fileext = '.lib'
elif option[0] == 'spice' or option[0] == 'spi':
fileext = '.spice' if not ef_format else '.spi'
elif option[0] == 'techlef':
fileext = '.lef'
else:
fileext = '.' + option[0]
if newname:
if os.path.splitext(newname)[1] == '':
newname = newname + fileext
if len(liblist) == 1:
destfile = newname
elif newname.startswith('*.'):
destfile = os.path.splitext(libfile)[0] + newname[1:]
elif newname.startswith('.'):
destfile = os.path.splitext(libfile)[0] + newname
else:
if not do_compile and not do_compile_only:
print('Error: rename specified but more than one file found!')
destfile = libfile
else:
destfile = libfile
targname = destlibdir + destpath + '/' + destfile
# NOTE: When using "up" with link_from, could just make
# destpath itself a symbolic link; this way is more flexible
# but adds one symbolic link per file.
if destpath != '':
if not os.path.isdir(destlibdir + destpath):
os.makedirs(destlibdir + destpath, exist_ok=True)
# Remove any existing file
if os.path.isfile(targname):
if not dontcopy:
os.remove(targname)
elif os.path.isdir(targname):
if not dontcopy:
shutil.rmtree(targname)
# Stage items for later removal if requested
if do_remove:
removelist.append(targname)
# NOTE: Diagnostic, probably much too much output.
print(' Install:' + libname + ' to ' + targname)
if not os.path.isdir(libname):
if not dontcopy:
if do_collate:
if libname not in splitfiles:
shutil.copy(libname, targname)
else:
allparts = splitfiles[libname]
with open(targname, 'wb') as afd:
for filepart in allparts:
with open(filepart, 'rb') as fd:
shutil.copyfileobj(fd, afd)
else:
shutil.copy(libname, targname)
else:
if not dontcopy:
shutil.copytree(libname, targname)
# File filtering options: Two options 'stub' and 'nospec' are
# handled by scripts in ../common/. Custom filters can also be
# specified.
local_filter_scripts = filter_scripts[:]
if option[0] == 'verilog':
# Internally handle syntactical issues with verilog and iverilog
vfilter(targname)
if do_remove_spec:
local_filter_scripts.append(scriptdir + '/remove_specify.py')
elif option[0] == 'cdl' or option[0] == 'spi' or option[0] == 'spice':
if do_stub:
local_filter_scripts.append(scriptdir + '/makestub.py')
for filter_script in local_filter_scripts:
# Apply filter script to all files in the target directory
tfilter(targname, filter_script, ef_format)
destfilelist.append(os.path.split(targname)[1])
# If headerfile is non-null, then copy this file, too. Do not add
# it to "destfilelist", as it is handled separately. Recast
# headerfile as the name of the file without the path.
if headerfile:
headerpath = substitute(sourcedir + '/' + headerfile, library[1])
headerlist = glob.glob(headerpath)
if len(headerlist) == 1:
libname = headerlist[0]
destfile = os.path.split(libname)[1]
targname = destlibdir + destpath + '/' + destfile
shutil.copy(libname, targname)
headerfile = destfile
# Add names from "include" list to destfilelist before writing
# filelist.txt for library file compiling.
for incname in includelist:
if '/' in incname:
# Names come from files in a path that is not the source
incfiles = os.listdir(os.path.split(incname)[0])
pattern = os.path.split(incname)[1]
destfilelist.extend(fnmatch.filter(incfiles, pattern))
else:
destfilelist.extend(fnmatch.filter(liblistnames, incname))
if sortscript:
with open(destlibdir + '/filelist.txt', 'w') as ofile:
for destfile in destfilelist:
print(destfile, file=ofile)
if os.path.isfile(sortscript):
print('Diagnostic: Sorting files with ' + sortscript)
subprocess.run(
[sortscript, destlibdir],
stdout = subprocess.DEVNULL,
stderr = subprocess.DEVNULL,
check = True,
)
if do_compile == True or do_compile_only == True:
# NOTE: The purpose of "rename" is to put a destlib-named
# library elsewhere so that it can be merged with another
# library into a compiled <destlib>.<ext> on another pass.
compname = destlib
# To do: Make this compatible with linking from another PDK.
if option[0] == 'verilog':
# If there is not a single file with all verilog cells in it,
# then compile one, because one does not want to have to have
# an include line for every single cell used in a design.
create_verilog_library(destlibdir, compname, do_compile_only, do_stub, excludelist)
elif option[0] == 'gds' and have_mag_8_2:
# If there is not a single file with all GDS cells in it,
# then compile one.
# Link to the PDK magic startup file from the target directory
startup_script = targetdir + mag_current + pdkname + '-F.magicrc'
if not os.path.isfile(startup_script):
startup_script = targetdir + mag_current + pdkname + '.magicrc'
create_gds_library(destlibdir, compname, startup_script, do_compile_only, excludelist)
elif option[0] == 'liberty' or option[0] == 'lib':
# If there is not a single file with all liberty cells in it,
# then compile one, because one does not want to have to have
# an include line for every single cell used in a design.
create_lib_library(destlibdir, compname, do_compile_only, excludelist, headerfile)
elif option[0] == 'spice' or option[0] == 'spi':
# If there is not a single file with all SPICE subcircuits in it,
# then compile one, because one does not want to have to have
# an include line for every single cell used in a design.
spiext = '.spice' if not ef_format else '.spi'
create_spice_library(destlibdir, compname, spiext, do_compile_only, do_stub, excludelist)
elif option[0] == 'cdl':
# If there is not a single file with all CDL subcircuits in it,
# then compile one, because one does not want to have to have
# an include line for every single cell used in a design.
create_spice_library(destlibdir, compname, '.cdl', do_compile_only, do_stub, excludelist)
elif option[0] == 'lef':
# If there is not a single file with all LEF cells in it,
# then compile one, because one does not want to have to have
# an include line for every single cell used in a design.
if not have_lefanno:
create_lef_library(destlibdir, compname, do_compile_only, excludelist)
if do_compile_only == True:
if newname and targname:
if os.path.isfile(targname):
if have_lefanno:
# If the original source is a single file
# but is used for annotation, then save the
# file name and delete it just before writing
# the LEF library
lef_savelibname = targname
else:
os.remove(targname)
# "rename" with "compile" or "compile-only": Change the name
# of the compiled file.
if newname:
print(' Renaming ' + compname + fileext + ' to ' + newname)
origname = destlibdir + '/' + compname + fileext
targrename = destlibdir + destpath + '/' + newname
if os.path.isfile(origname):
os.rename(origname, targrename)
# If "filelist.txt" was created, remove it
if sortscript:
if os.path.isfile(destlibdir + '/filelist.txt'):
os.remove(destlibdir + '/filelist.txt')
# Find any libraries/options marked as "privileged" (or "private") and
# move the files from libs.tech or libs.ref to libs.priv, leaving a
# symbolic link in the original location. Do this during the initial
# install so that options following in the list can add files to the
# non-privileged equivalent directory path.
if 'priv' in option or 'privileged' in option or 'private' in option:
# Diagnostic
print("Install option: " + str(option[0]))
if ef_format == True:
os.makedirs(targetdir + '/libs.priv', exist_ok=True)
for library in libraries:
if len(library) == 3:
destlib = library[2]
else:
destlib = library[1]
if ef_format:
srclibdir = targetdir + '/libs.ref/' + option[0] + '/' + destlib
destlibdir = targetdir + '/libs.priv/' + option[0] + '/' + destlib
else:
srclibdir = targetdir + '/libs.ref/' + destlib + '/' + option[0]
destlibdir = targetdir + '/libs.priv/' + destlib + '/' + option[0]
if not os.path.exists(destlibdir):
os.makedirs(destlibdir)
print('Moving files in ' + srclibdir + ' to privileged space.')
filelist = os.listdir(srclibdir)
for file in filelist:
srcfile = srclibdir + '/' + file
destfile = destlibdir + '/' + file
if os.path.isfile(destfile):
os.remove(destfile)
elif os.path.isdir(destfile):
shutil.rmtree(destfile)
if os.path.isfile(srcfile):
shutil.copy(srcfile, destfile)
os.remove(srcfile)
else:
shutil.copytree(srcfile, destfile)
shutil.rmtree(srcfile)
print("Completed installation of vendor files.")
#----------------------------------------------------------------
# Installation part 2: Generate derived file formats
#----------------------------------------------------------------
# Now for the harder part. If GDS and/or LEF databases were specified,
# then migrate them to magic (.mag files in layout/ or abstract/).
ignorelist = []
tclscript = None
do_cdl_scaleu = False
no_cdl_convert = False
no_gds_convert = False
no_lef_convert = False
do_parasitics = False
no_extract = False
cdl_compile_only = False
lef_compile = False
lef_compile_only = False
lefopts = None
cdl_exclude = []
lef_exclude = []
gds_exclude = []
spice_exclude = []
verilog_exclude = []
cdl_reflib = '/libs.ref/'
gds_reflib = '/libs.ref/'
lef_reflib = '/libs.ref/'
for option in optionlist[:]:
if option[0] == 'cdl':
# Option 'scaleu' is a standalone keyword
do_cdl_scaleu = 'scaleu' in option
# Option 'ignore' has arguments after '='
for item in option:
if item.split('=')[0] == 'ignore':
ignorelist = item.split('=')[1].split(',')
elif option[0] == 'gds':
for item in option:
if item.split('=')[0] == 'options':
tclscript = item.split('=')[1]
tcllines = []
print('Adding Tcl script options from file ' + tclscript)
# Option 'noconvert' is a standalone keyword.
if 'noconvert' in option:
if option[0] == 'cdl':
no_cdl_convert = True
elif option[0] == 'gds':
no_gds_convert = True
elif option[0] == 'lef':
no_lef_convert = True
if 'dorcx' in option:
if option[0] == 'gds':
do_parasitics = True
if 'noextract' in option:
if option[0] == 'gds':
no_extract = True
# Option 'privileged' is a standalone keyword.
if 'priv' in option or 'privileged' in option or 'private' in option:
if option[0] == 'cdl':
cdl_reflib = '/libs.priv/'
elif option[0] == 'gds':
gds_reflib = '/libs.priv/'
elif option[0] == 'lef':
lef_reflib = '/libs.priv/'
# If CDL is marked 'compile-only' then CDL should only convert the
# compiled file to SPICE if conversion is needed. If LEF is marked
# 'compile' or 'compile-only' in annotate mode, then create a LEF
# library from magic LEF output.
if 'compile-only' in option:
if option[0] == 'cdl':
cdl_compile_only = True
elif option[0] == 'lef':
lef_compile_only = True
elif 'compile' in option:
if option[0] == 'lef':
lef_compile = True
# Find exclude list for any option
for item in option:
if item.split('=')[0] == 'exclude':
exclude_list = item.split('=')[1].split(',')
if option[0] == 'cdl':
cdl_exclude = exclude_list
elif option[0] == 'lef':
lef_exclude = exclude_list
elif option[0] == 'gds':
gds_exclude = exclude_list
elif option[0] == 'spi' or option[0] == 'spice':
spice_exclude = exclude_list
elif option[0] == 'verilog':
verilog_exclude = exclude_list
# Find options list for "lef write"
for item in option:
if item.split('=')[0] == 'lefopts':
if option[0] == 'lef':
lefopts = item.split('=')[1].strip('"')
devlist = []
pdklibrary = None
if tclscript:
# If tclscript is a file, then read it. Otherwise, assume
# that the option contents should be inserted verbatim.
if os.path.isfile(tclscript):
with open(tclscript, 'r') as ifile:
tcllines = ifile.read().splitlines()
else:
tcllines = list(tclscript)
if have_gds and not no_gds_convert:
print("Migrating GDS files to layout.")
if ef_format:
destdir = targetdir + gds_reflib + 'mag'
srcdir = targetdir + gds_reflib + 'gds'
vdir = targetdir + '/libs.ref/' + 'verilog'
cdir = targetdir + cdl_reflib + 'cdl'
sdir = targetdir + cdl_reflib + 'spi'
os.makedirs(destdir, exist_ok=True)
# For each library, create the library subdirectory
for library in libraries:
if len(library) == 3:
destlib = library[2]
else:
destlib = library[1]
if ef_format:
destlibdir = destdir + '/' + destlib
srclibdir = srcdir + '/' + destlib
vlibdir = vdir + '/' + destlib
clibdir = cdir + '/' + destlib
slibdir = sdir + '/' + destlib
else:
destdir = targetdir + gds_reflib + destlib + '/mag'
srcdir = targetdir + gds_reflib + destlib + '/gds'
vdir = targetdir + '/libs.ref/' + destlib + '/verilog'
cdir = targetdir + cdl_reflib + destlib + '/cdl'
sdir = targetdir + cdl_reflib + destlib + '/spice'
destlibdir = destdir
srclibdir = srcdir
vlibdir = vdir
clibdir = cdir
slibdir = sdir
os.makedirs(destlibdir, exist_ok=True)
# For primitive devices, check the PDK script and find the name
# of the library and get a list of supported devices.
if library[0] == 'primitive':
pdkscript = targetdir + mag_current + pdkname + '.tcl'
print('Searching for supported devices in PDK script ' + pdkscript + '.')
if os.path.isfile(pdkscript):
librex = re.compile('^[ \t]*set[ \t]+PDKNAMESPACE[ \t]+([^ \t]+)$')
devrex = re.compile('^[ \t]*proc[ \t]+([^ :\t]+)::([^ \t_]+)_defaults')
fixrex = re.compile('^[ \t]*return[ \t]+\[([^ :\t]+)::fixed_draw[ \t]+([^ \t]+)[ \t]+')
devlist = []
fixedlist = []
with open(pdkscript, 'r') as ifile:
scripttext = ifile.read().splitlines()
for line in scripttext:
lmatch = librex.match(line)
if lmatch:
pdklibrary = lmatch.group(1)
dmatch = devrex.match(line)
if dmatch:
if dmatch.group(1) == pdklibrary:
devlist.append(dmatch.group(2))
fmatch = fixrex.match(line)
if fmatch:
if fmatch.group(1) == pdklibrary:
fixedlist.append(fmatch.group(2))
# Diagnostic
print("PDK library is " + str(pdklibrary))
# Link to the PDK magic startup file from the target directory
# If there is no -F version then look for one without -F (open source PDK)
startup_script = targetdir + mag_current + pdkname + '-F.magicrc'
if not os.path.isfile(startup_script):
startup_script = targetdir + mag_current + pdkname + '.magicrc'
if have_mag_8_2 and os.path.isfile(startup_script):
# If the symbolic link exists, remove it.
if os.path.isfile(destlibdir + '/.magicrc'):
os.remove(destlibdir + '/.magicrc')
os.symlink(startup_script, destlibdir + '/.magicrc')
# Find GDS file names in the source
print('Getting GDS file list from ' + srclibdir + '.')
gdsfilesraw = os.listdir(srclibdir)
gdsfiles = []
for gdsfile in gdsfilesraw:
gdsext = os.path.splitext(gdsfile)[1].lower()
if gdsext == '.gds' or gdsext == '.gdsii' or gdsext == '.gds2':
gdsfiles.append(gdsfile)
# Create exclude list with glob-style matching using fnmatch
if len(gdsfiles) > 0:
gdsnames = list(os.path.split(item)[1] for item in gdsfiles)
notgdsnames = []
for exclude in gds_exclude:
notgdsnames.extend(fnmatch.filter(gdsnames, exclude))
# Apply exclude list
if len(notgdsnames) > 0:
for file in gdsfiles[:]:
if os.path.split(file)[1] in notgdsnames:
gdsfiles.remove(file)
# Generate a script called "generate_magic.tcl" and leave it in
# the target directory. Use it as input to magic to create the
# .mag files from the database.
print('Creating magic generation script to generate magic database files.')
with open(destlibdir + '/generate_magic.tcl', 'w') as ofile:
print('#!/usr/bin/env wish', file=ofile)
print('#--------------------------------------------', file=ofile)
print('# Script to generate .mag files from .gds ', file=ofile)
print('#--------------------------------------------', file=ofile)
print('crashbackups stop', file=ofile)
print('drc off', file=ofile)
print('locking off', file=ofile)
if do_timestamp and have_mag_8_3_261:
print('gds datestamp ' + str(timestamp_value), file=ofile)
print('gds readonly true', file=ofile)
print('gds drccheck false', file=ofile)
print('gds flatten true', file=ofile)
print('gds rescale false', file=ofile)
print('tech unlock *', file=ofile)
# Add custom Tcl script lines before "gds read".
if tclscript:
for line in tcllines:
print(line, file=ofile)
for gdsfile in gdsfiles:
# Note: DO NOT use a relative path here.
print('gds read ' + srclibdir + '/' + gdsfile, file=ofile)
# Make sure properties include the Tcl generated cell
# information from the PDK script
if pdklibrary:
tclfixedlist = '{' + ' '.join(fixedlist) + '}'
print('set devlist ' + tclfixedlist, file=ofile)
print('set topcell [lindex [cellname list top] 0]',
file=ofile)
print('foreach cellname $devlist {', file=ofile)
print(' load $cellname', file=ofile)
print(' property gencell $cellname', file=ofile)
print(' property parameter m=1', file=ofile)
print(' property library ' + pdklibrary, file=ofile)
print('}', file=ofile)
print('load $topcell', file=ofile)
else:
# Use LEF files to set the port properties
if have_lefanno or have_lef:
lefdirname = 'lef'
# Find LEF file names in the source
if ef_format:
lefsrcdir = targetdir + lef_reflib + lefdirname
lefsrclibdir = lefsrcdir + '/' + destlib
else:
lefsrcdir = targetdir + lef_reflib + destlib + '/' + lefdirname
lefsrclibdir = lefsrcdir
leffiles = os.listdir(lefsrclibdir)
leffiles = list(item for item in leffiles if os.path.splitext(item)[1] == '.lef')
if len(leffiles) > 0:
lefnames = list(os.path.split(item)[1] for item in leffiles)
notlefnames = []
for exclude in lef_exclude:
notlefnames.extend(fnmatch.filter(lefnames, exclude))
# Apply exclude list
if len(notlefnames) > 0:
for file in leffiles[:]:
if os.path.split(file)[1] in notlefnames:
leffiles.remove(file)
if len(leffiles) > 0:
print('puts stdout "Annotating cells from LEF"', file=ofile)
for leffile in leffiles:
print('lef read ' + lefsrclibdir + '/' + leffile, file=ofile)
# Use CDL or SPICE netlists to set the port order
if have_cdl or have_spice:
if have_cdl:
netdir = clibdir
else:
netdir = slibdir
# Find CDL/SPICE file names in the source
# Ignore "sources.txt" if it is in the list.
netfiles = os.listdir(netdir)
print('puts stdout "Annotating cells from CDL/SPICE"',
file=ofile)
for netfile in netfiles:
if os.path.split(netfile)[1] != 'sources.txt':
print('catch {readspice ' + netdir + '/' + netfile
+ '}', file=ofile)
# print('cellname delete \(UNNAMED\)', file=ofile)
print('puts stdout "Writing all magic database files"', file=ofile)
print('writeall force', file=ofile)
leffiles = []
lefmacros = []
if have_lefanno:
# Find LEF file names in the source
if ef_format:
lefsrcdir = targetdir + lef_reflib + 'lef'
lefsrclibdir = lefsrcdir + '/' + destlib
else:
lefsrcdir = targetdir + lef_reflib + destlib + '/lef'
lefsrclibdir = lefsrcdir
leffiles = os.listdir(lefsrclibdir)
leffiles = list(item for item in leffiles if os.path.splitext(item)[1] == '.lef')
# Create exclude list with glob-style matching using fnmatch
if len(leffiles) > 0:
lefnames = list(os.path.split(item)[1] for item in leffiles)
notlefnames = []
for exclude in lef_exclude:
notlefnames.extend(fnmatch.filter(lefnames, exclude))
# Apply exclude list
if len(notlefnames) > 0:
for file in leffiles[:]:
if os.path.split(file)[1] in notlefnames:
leffiles.remove(file)
# Get list of abstract views to make from LEF macros
# (Note: exclude list can only contain the file being
# read, not individual macro names in the file; might
# need some additional feature to accommodate this.)
for leffile in leffiles:
with open(lefsrclibdir + '/' + leffile, 'r') as ifile:
ltext = ifile.read()
llines = ltext.splitlines()
for lline in llines:
ltok = re.split(' |\t|\(', lline)
if ltok[0] == 'MACRO':
lefmacros.append(ltok[1])
elif have_lef:
# Nothing to do; LEF macros were already installed.
pass
elif have_verilog and os.path.isdir(vlibdir):
# Get list of abstract views to make from verilog modules
# (NOTE: no way to apply exclude list here!)
vfiles = os.listdir(vlibdir)
vfiles = list(item for item in vfiles if os.path.splitext(item)[1] == '.v')
# Create exclude list with glob-style matching using fnmatch
if len(vfiles) > 0:
vnames = list(os.path.split(item)[1] for item in vfiles)
notvnames = []
for exclude in verilog_exclude:
notvnames.extend(fnmatch.filter(vnames, exclude))
# Apply exclude list
if len(notvnames) > 0:
for file in vfiles[:]:
if os.path.split(file)[1] in notvnames:
vfiles.remove(file)
for vfile in vfiles:
with open(vlibdir + '/' + vfile, 'r') as ifile:
vtext = ifile.read()
vlines = vtext.splitlines()
for vline in vlines:
vtok = re.split(' |\t|\(', vline)
try:
if vtok[0] == 'module':
if vtok[1] not in lefmacros:
lefmacros.append(vtok[1])
except:
pass
elif have_cdl and os.path.isdir(clibdir):
# Get list of abstract views to make from CDL subcircuits
cfiles = os.listdir(clibdir)
cfiles = list(item for item in cfiles if os.path.splitext(item)[1] == '.cdl')
# Create exclude list with glob-style matching using fnmatch
if len(cfiles) > 0:
cnames = list(os.path.split(item)[1] for item in cfiles)
notcnames = []
for exclude in cdl_exclude:
notcnames.extend(fnmatch.filter(cnames, exclude))
# Apply exclude list
if len(notcnames) > 0:
for file in cfiles[:]:
if os.path.split(file)[1] in notcnames:
cfiles.remove(file)
for cfile in cfiles:
with open(clibdir + '/' + cfile, 'r') as ifile:
ctext = ifile.read()
clines = ctext.splitlines()
for cline in clines:
ctok = cline.split()
try:
if ctok[0].lower() == '.subckt':
if ctok[1] not in lefmacros:
lefmacros.append(ctok[1])
except:
pass
elif have_spice and os.path.isdir(slibdir):
# Get list of abstract views to make from SPICE subcircuits
sfiles = os.listdir(slibdir)
sfiles = list(item for item in sfiles)
# Create exclude list with glob-style matching using fnmatch
if len(sfiles) > 0:
snames = list(os.path.split(item)[1] for item in sfiles)
notsnames = []
for exclude in spice_exclude:
notsnames.extend(fnmatch.filter(snames, exclude))
# Apply exclude list
if len(notsnames) > 0:
for file in sfiles[:]:
if os.path.split(file)[1] in notsnames:
sfiles.remove(file)
for sfile in sfiles:
with open(slibdir + '/' + sfile, 'r') as ifile:
stext = ifile.read()
slines = stext.splitlines()
for sline in slines:
stok = sline.split()
try:
if stok[0].lower() == '.subckt':
if stok[1] not in lefmacros:
lefmacros.append(stok[1])
except:
pass
if not lefmacros:
print('No source for abstract views: Abstract views not made.')
elif have_lefanno or not have_lef:
# This library has a GDS database but no LEF database. Use
# magic to create abstract views of the GDS cells. If
# option "annotate" is given, then read the LEF file after
# loading the database file to annotate the cell with
# information from the LEF file. This usually indicates
# that the LEF file has some weird definition of obstruction
# layers and we want to normalize them by using magic's LEF
# write procedure, but we still need the pin use and class
# information from the LEF file, and maybe the bounding box.
# For annotation, the LEF file output will overwrite the
# original source LEF file.
lefdest = lefsrclibdir + '/' if have_lefanno else ''
# Delete the original files in case the naming is different
for leffile in leffiles:
print('file delete ' + lefsrclibdir + '/' + leffile, file=ofile)
for lefmacro in lefmacros:
print('if {[cellname list exists ' + lefmacro + '] != 0} {', file=ofile)
print(' load ' + lefmacro, file=ofile)
if lefopts:
print(' lef write ' + lefdest + lefmacro + ' ' + lefopts, file=ofile)
else:
print(' lef write ' + lefdest + lefmacro, file=ofile)
print('}', file=ofile)
print('puts stdout "Done."', file=ofile)
print('quit -noprompt', file=ofile)
print('Running magic to create magic database files.')
sys.stdout.flush()
# Run magic to read in the GDS file and write out magic databases.
with open(destlibdir + '/generate_magic.tcl', 'r') as ifile:
subprocess_run(
'magic',
['magic', '-dnull', '-noconsole'],
stdin = ifile,
cwd = destlibdir,
)
# Set have_lef now that LEF files were made, so they
# can be used to generate the maglef/ databases.
have_lef = True
elif not have_mag_8_2:
print('The installer is not able to run magic.')
else:
print("Master PDK magic startup file not found. Did you install")
print("PDK tech files before PDK vendor files?")
if have_lefanno:
# LEF files were used for annotation. If "compile" or "compile-only"
# was also passed as an option, then build the LEF library now from
# the LEF output from magic.
print("Compiling LEF library from magic output.")
# If a single vendor library file was used for annotation, remove it now.
if lef_savelibname:
if os.path.isfile(lef_savelibname):
os.remove(lef_savelibname)
if lef_compile or lef_compile_only:
create_lef_library(lefsrclibdir, destlib, lef_compile_only, lef_exclude)
if have_lef and not no_lef_convert:
print("Migrating LEF files to layout.")
if ef_format:
destdir = targetdir + '/libs.ref/' + 'maglef'
srcdir = targetdir + lef_reflib + 'lef'
magdir = targetdir + gds_reflib + 'mag'
cdldir = targetdir + cdl_reflib + 'cdl'
os.makedirs(destdir, exist_ok=True)
# For each library, create the library subdirectory
for library in libraries:
if len(library) == 3:
destlib = library[2]
else:
destlib = library[1]
if ef_format:
destlibdir = destdir + '/' + destlib
srclibdir = srcdir + '/' + destlib
maglibdir = magdir + '/' + destlib
cdllibdir = cdldir + '/' + destlib
clibdir = cdir + '/' + destlib
slibdir = sdir + '/' + destlib
else:
destdir = targetdir + '/libs.ref/' + destlib + '/maglef'
srcdir = targetdir + lef_reflib + destlib + '/lef'
magdir = targetdir + gds_reflib + destlib + '/mag'
cdldir = targetdir + cdl_reflib + destlib + '/cdl'
cdir = targetdir + cdl_reflib + destlib + '/cdl'
sdir = targetdir + cdl_reflib + destlib + '/spice'
destlibdir = destdir
srclibdir = srcdir
maglibdir = magdir
cdllibdir = cdldir
clibdir = cdir
slibdir = sdir
os.makedirs(destlibdir, exist_ok=True)
# Link to the PDK magic startup file from the target directory
startup_script = targetdir + mag_current + pdkname + '-F.magicrc'
if not os.path.isfile(startup_script):
startup_script = targetdir + mag_current + pdkname + '.magicrc'
if have_mag_8_2 and os.path.isfile(startup_script):
# If the symbolic link exists, remove it.
if os.path.isfile(destlibdir + '/.magicrc'):
os.remove(destlibdir + '/.magicrc')
os.symlink(startup_script, destlibdir + '/.magicrc')
# Find LEF file names in the source
leffiles = []
if os.path.isdir(srclibdir):
leffiles = os.listdir(srclibdir)
leffiles = list(item for item in leffiles if os.path.splitext(item)[1].lower() == '.lef')
# Get list of abstract views to make from LEF macros
lefmacros = []
err_no_macros = False
for leffile in leffiles:
with open(srclibdir + '/' + leffile, 'r') as ifile:
ltext = ifile.read()
llines = ltext.splitlines()
for lline in llines:
ltok = re.split(' |\t|\(', lline)
if ltok[0] == 'MACRO':
lefmacros.append(ltok[1])
# Create exclude list with glob-style matching using fnmatch
if len(lefmacros) > 0:
lefnames = list(os.path.split(item)[1] for item in lefmacros)
notlefnames = []
for exclude in lef_exclude:
notlefnames.extend(fnmatch.filter(lefnames, exclude))
# Apply exclude list
if len(notlefnames) > 0:
for file in lefmacros[:]:
if os.path.split(file)[1] in notlefnames:
lefmacros.remove(file)
if len(leffiles) == 0:
print('Warning: No LEF files found in ' + srclibdir)
continue
print('Generating conversion script to create magic databases from LEF')
# Generate a script called "generate_magic.tcl" and leave it in
# the target directory. Use it as input to magic to create the
# .mag files from the database.
with open(destlibdir + '/generate_magic.tcl', 'w') as ofile:
print('#!/usr/bin/env wish', file=ofile)
print('#--------------------------------------------', file=ofile)
print('# Script to generate .mag files from .lef ', file=ofile)
print('#--------------------------------------------', file=ofile)
print('tech unlock *', file=ofile)
# If there are devices in the LEF file that come from the
# PDK library, then copy this list into the script.
if pdklibrary:
shortdevlist = []
for macro in lefmacros:
if macro in devlist:
shortdevlist.append(macro)
tcldevlist = '{' + ' '.join(shortdevlist) + '}'
print('set devlist ' + tcldevlist, file=ofile)
# Force the abstract view timestamps to match the full views
if do_timestamp and have_mag_8_3_261:
print('lef datestamp ' + str(timestamp_value), file=ofile)
for leffile in leffiles:
print('lef read ' + srclibdir + '/' + leffile, file=ofile)
# Use CDL or SPICE netlists to make sure that ports are
# present, and to set the port order
if have_cdl or have_spice:
if have_cdl:
netdir = clibdir
else:
netdir = slibdir
# Find CDL/SPICE file names in the source
# Ignore "sources.txt" if it is in the list.
netfiles = os.listdir(netdir)
print('puts stdout "Annotating cells from CDL/SPICE"',
file=ofile)
for netfile in netfiles:
if os.path.split(netfile)[1] != 'sources.txt':
print('catch {readspice ' + netdir + '/' + netfile
+ '}', file=ofile)
for lefmacro in lefmacros:
if pdklibrary and lefmacro in shortdevlist:
print('set cellname ' + lefmacro, file=ofile)
print('if {[lsearch $devlist $cellname] >= 0} {',
file=ofile)
print(' load $cellname', file=ofile)
print(' property gencell $cellname', file=ofile)
print(' property parameter m=1', file=ofile)
print(' property library ' + pdklibrary, file=ofile)
print('}', file=ofile)
# Load one of the LEF files so that the default (UNNAMED) cell
# is not loaded, then delete (UNNAMED) so it doesn't generate
# an error message.
if len(lefmacros) > 0:
print('load ' + lefmacros[0], file=ofile)
# print('cellname delete \(UNNAMED\)', file=ofile)
else:
err_no_macros = True
print('writeall force', file=ofile)
print('puts stdout "Done."', file=ofile)
print('quit -noprompt', file=ofile)
if err_no_macros == True:
print('Warning: No LEF macros were defined.')
print('Running magic to create magic databases from LEF')
sys.stdout.flush()
# Run magic to read in the LEF file and write out magic databases.
with open(destlibdir + '/generate_magic.tcl', 'r') as ifile:
subprocess_run(
'magic',
['magic', '-dnull', '-noconsole'],
stdin = ifile,
cwd = destlibdir,
)
# Now list all the .mag files generated, and for each, read the
# corresponding file from the mag/ directory, pull the GDS file
# properties, and add those properties to the maglef view. Also
# read the CDL (or SPICE) netlist, read the ports, and rewrite
# the port order in the mag and maglef file accordingly.
# Diagnostic
print('Annotating files in ' + destlibdir)
sys.stdout.flush()
magfiles = os.listdir(destlibdir)
magfiles = list(item for item in magfiles if os.path.splitext(item)[1] == '.mag')
for magroot in magfiles:
magname = os.path.splitext(magroot)[0]
magfile = maglibdir + '/' + magroot
magleffile = destlibdir + '/' + magroot
prop_lines = get_gds_properties(magfile)
# Make sure properties include the Tcl generated cell
# information from the PDK script
prop_gencell = []
if pdklibrary:
if magname in fixedlist:
prop_gencell.append('gencell ' + magname)
prop_gencell.append('library ' + pdklibrary)
prop_gencell.append('parameter m=1')
nprops = len(prop_lines) + len(prop_gencell)
cdlfile = cdllibdir + '/' + magname + '.cdl'
if os.path.exists(cdlfile):
cdlfiles = [cdlfile]
else:
# Assume there is at least one file with all cell subcircuits
# in it.
try:
cdlfiles = glob.glob(cdllibdir + '/*.cdl')
cdlfiles = natural_sort.natural_sort(cdlfiles)
except:
pass
if len(cdlfiles) > 0:
for cdlfile in cdlfiles:
port_dict = get_subckt_ports(cdlfile, magname)
if port_dict != {}:
break
else:
port_dict = {}
if port_dict == {}:
print('No CDL file contains ' + destlib + ' device ' + magname)
cdlfile = None
# To be done: If destlib is 'primitive', then look in
# SPICE models for port order.
if destlib == 'primitive':
print('Fix me: Need to look in SPICE models!')
proprex = re.compile('<< properties >>')
endrex = re.compile('<< end >>')
rlabrex = re.compile('rlabel[ \t]+[^ \t]+[ \t]+[^ \t]+[ \t]+[^ \t]+[ \t]+[^ \t]+[ \t]+[^ \t]+[ \t]+[^ \t]+[ \t]+([^ \t]+)')
flabrex = re.compile('flabel[ \t]+.*[ \t]+([^ \t]+)[ \t]*')
portrex = re.compile('port[ \t]+([^ \t]+)[ \t]+(.*)')
gcellrex = re.compile('string gencell')
portnum = -1
with open(magleffile, 'r') as ifile:
magtext = ifile.read().splitlines()
with open(magleffile, 'w') as ofile:
has_props = False
is_gencell = False
for line in magtext:
tmatch = portrex.match(line)
if tmatch:
if portnum >= 0:
line = 'port ' + str(portnum) + ' ' + tmatch.group(2)
else:
line = 'port ' + tmatch.group(1) + ' ' + tmatch.group(2)
ematch = endrex.match(line)
if ematch and nprops > 0:
if not has_props:
print('<< properties >>', file=ofile)
if not is_gencell:
for prop in prop_gencell:
print('string ' + prop, file=ofile)
for prop in prop_lines:
print('string ' + prop, file=ofile)
print(line, file=ofile)
pmatch = proprex.match(line)
if pmatch:
has_props = True
gmatch = gcellrex.match(line)
if gmatch:
is_gencell = True
lmatch = flabrex.match(line)
if not lmatch:
lmatch = rlabrex.match(line)
if lmatch:
labname = lmatch.group(1).lower()
try:
portnum = port_dict[labname]
except:
portnum = -1
if os.path.exists(magfile):
with open(magfile, 'r') as ifile:
magtext = ifile.read().splitlines()
with open(magfile, 'w') as ofile:
for line in magtext:
tmatch = portrex.match(line)
if tmatch:
if portnum >= 0:
line = 'port ' + str(portnum) + ' ' + tmatch.group(2)
else:
line = 'port ' + tmatch.group(1) + ' ' + tmatch.group(2)
ematch = endrex.match(line)
print(line, file=ofile)
lmatch = flabrex.match(line)
if not lmatch:
lmatch = rlabrex.match(line)
if lmatch:
labname = lmatch.group(1).lower()
try:
portnum = port_dict[labname]
except:
portnum = -1
elif os.path.splitext(magfile)[1] == '.mag':
# NOTE: Possibly this means the GDS cell has a different name.
print('Error: No file ' + magfile + '. Why is it in maglef???')
elif not have_mag_8_2:
print('The installer is not able to run magic.')
else:
print("Master PDK magic startup file not found. Did you install")
print("PDK tech files before PDK vendor files?")
# If SPICE or CDL databases were specified, then convert them to
# a form that can be used by ngspice, using the cdl2spi.py script
if have_spice:
if ef_format:
if not os.path.isdir(targetdir + cdl_reflib + 'spi'):
os.makedirs(targetdir + cdl_reflib + 'spi', exist_ok=True)
elif have_cdl and not no_cdl_convert:
if ef_format:
if not os.path.isdir(targetdir + cdl_reflib + 'spi'):
os.makedirs(targetdir + cdl_reflib + 'spi', exist_ok=True)
print("Migrating CDL netlists to SPICE.")
sys.stdout.flush()
if ef_format:
destdir = targetdir + cdl_reflib + 'spi'
srcdir = targetdir + cdl_reflib + 'cdl'
os.makedirs(destdir, exist_ok=True)
# For each library, create the library subdirectory
for library in libraries:
if len(library) == 3:
destlib = library[2]
else:
destlib = library[1]
if ef_format:
destlibdir = destdir + '/' + destlib
srclibdir = srcdir + '/' + destlib
else:
destdir = targetdir + cdl_reflib + destlib + '/spice'
srcdir = targetdir + cdl_reflib + destlib + '/cdl'
destlibdir = destdir
srclibdir = srcdir
os.makedirs(destlibdir, exist_ok=True)
# Find CDL file names in the source
# If CDL is marked compile-only then ONLY convert <distdir>.cdl
if cdl_compile_only:
alllibname = destlibdir + '/' + destlib + '.cdl'
if not os.path.exists(alllibname):
cdl_compile_only = False
else:
cdlfiles = [alllibname]
if not cdl_compile_only:
cdlfiles = os.listdir(srclibdir)
cdlfiles = list(item for item in cdlfiles if os.path.splitext(item)[1].lower() == '.cdl')
# The directory with scripts should be in ../common with respect
# to the Makefile that determines the cwd.
# Run cdl2spi.py script to read in the CDL file and write out SPICE
for cdlfile in cdlfiles:
if ef_format:
spiname = os.path.splitext(cdlfile)[0] + '.spi'
else:
spiname = os.path.splitext(cdlfile)[0] + '.spice'
procopts = [scriptdir + '/cdl2spi.py', srclibdir + '/' + cdlfile, destlibdir + '/' + spiname]
if do_cdl_scaleu:
procopts.append('-dscale=u')
for item in ignorelist:
procopts.append('-ignore=' + item)
print('Running (in ' + destlibdir + '): ' + ' '.join(procopts))
subprocess_run('cdl2spi.py', procopts, cwd = destlibdir)
elif have_gds and not no_gds_convert and not no_extract:
# If neither SPICE nor CDL formats is available in the source, then
# read GDS; if the result has no ports, then read the corresponding
# LEF library to get port information. Then write out a SPICE netlist
# for the whole library. NOTE: If there is no CDL or SPICE source,
# then the port numbering is arbitrary, and becomes whatever the
# output of this script makes it.
if ef_format:
destdir = targetdir + cdl_reflib + 'spi'
srcdir = targetdir + gds_reflib + 'gds'
lefdir = targetdir + lef_reflib + 'lef'
cdldir = targetdir + cdl_reflib + 'cdl'
os.makedirs(destdir, exist_ok=True)
# For each library, create the library subdirectory
for library in libraries:
if len(library) == 3:
destlib = library[2]
else:
destlib = library[1]
if ef_format:
destlibdir = destdir + '/' + destlib
srclibdir = srcdir + '/' + destlib
leflibdir = lefdir + '/' + destlib
cdllibdir = cdldir + '/' + destlib
else:
destdir = targetdir + cdl_reflib + destlib + '/spice'
srcdir = targetdir + gds_reflib + destlib + '/gds'
lefdir = targetdir + lef_reflib + destlib + '/lef'
cdldir = targetdir + cdl_reflib + destlib + '/cdl'
destlibdir = destdir
srclibdir = srcdir
leflibdir = lefdir
cdllibdir = cdldir
os.makedirs(destlibdir, exist_ok=True)
# Link to the PDK magic startup file from the target directory
startup_script = targetdir + mag_current + pdkname + '-F.magicrc'
if not os.path.isfile(startup_script):
startup_script = targetdir + mag_current + pdkname + '.magicrc'
if os.path.isfile(startup_script):
# If the symbolic link exists, remove it.
if os.path.isfile(destlibdir + '/.magicrc'):
os.remove(destlibdir + '/.magicrc')
os.symlink(startup_script, destlibdir + '/.magicrc')
# Get the consolidated GDS library file, or a list of all GDS files
# if there is no single consolidated library
allgdslibname = srclibdir + '/' + destlib + '.gds'
if not os.path.isfile(allgdslibname):
glist = glob.glob(srclibdir + '/*.gds')
glist.extend(glob.glob(srclibdir + '/*.gdsii'))
glist.extend(glob.glob(srclibdir + '/*.gds2'))
glist = natural_sort.natural_sort(glist)
allleflibname = leflibdir + '/' + destlib + '.lef'
if not os.path.isfile(allleflibname):
llist = glob.glob(leflibdir + '/*.lef')
llist = natural_sort.natural_sort(llist)
if have_cdl and no_cdl_convert:
# CDL is not being converted directly to SPICE but it exists,
# and being the only source of pin order, it should be used
# for pin order annotation.
allcdllibname = cdllibdir + '/' + destlib + '.cdl'
if not os.path.isfile(allcdllibname):
clist = glob.glob(cdllibdir + '/*.cdl')
clist = natural_sort.natural_sort(clist)
print('Creating magic generation script to generate SPICE library.')
with open(destlibdir + '/generate_magic.tcl', 'w') as ofile:
print('#!/usr/bin/env wish', file=ofile)
print('#---------------------------------------------', file=ofile)
print('# Script to generate SPICE library from GDS ', file=ofile)
print('#---------------------------------------------', file=ofile)
print('drc off', file=ofile)
print('locking off', file=ofile)
print('gds readonly true', file=ofile)
print('gds flatten true', file=ofile)
print('gds rescale false', file=ofile)
print('tech unlock *', file=ofile)
# Add custom Tcl script lines before "gds read".
if tclscript:
for line in tcllines:
print(line, file=ofile)
if not os.path.isfile(allgdslibname):
for gdsfile in glist:
print('gds read ' + gdsfile, file=ofile)
else:
print('gds read ' + allgdslibname, file=ofile)
if not os.path.isfile(allleflibname):
# Annotate the cells with information from the LEF files
for leffile in llist:
print('lef read ' + leffile, file=ofile)
else:
print('lef read ' + allleflibname, file=ofile)
if have_cdl and no_cdl_convert:
if not os.path.isfile(allcdllibname):
# Annotate the cells with pin order from the CDL files
for cdlfile in clist:
print('catch {readspice ' + cdlfile + '}', file=ofile)
else:
print('catch {readspice ' + allcdllibname + '}', file=ofile)
# Load first file and remove the (UNNAMED) cell
if not os.path.isfile(allgdslibname):
print('load ' + os.path.splitext(glist[0])[0], file=ofile)
else:
gdslibroot = os.path.split(allgdslibname)[1]
print('load ' + os.path.splitext(gdslibroot)[0], file=ofile)
print('catch {cellname delete \(UNNAMED\)}', file=ofile)
print('ext2spice lvs', file=ofile)
# NOTE: Leaving "subcircuit top" as "auto" (default) can cause
# cells like decap that have no I/O to be output without a subcircuit
# wrapper. Also note that if this happens, it is an indication that
# power supplies have not been labeled as ports, which is harder to
# handle and should be fixed in the source.
print('ext2spice subcircuit top on', file=ofile)
# Use option "dorcx" if parasitics should be extracted.
# NOTE: Currently only does parasitic capacitance extraction.
if do_parasitics:
print('ext2spice cthresh 0.1', file=ofile)
if os.path.isfile(allgdslibname):
# Do not depend absolutely on the library having a top
# level cell, but query for it from inside magic
print('if {[cellname list exists ' + allgdslibname + ']} {',
file=ofile)
print(' select top cell', file=ofile)
print(' set glist [cellname list children]', file=ofile)
print('} else {', file=ofile)
print(' set glist [cellname list top]', file=ofile)
print('}', file=ofile)
else:
print('set glist [cellname list top]', file=ofile)
print('foreach cell $glist {', file=ofile)
print(' load $cell', file=ofile)
print(' puts stdout "Extracting cell $cell"', file=ofile)
print(' extract all', file=ofile)
print(' ext2spice', file=ofile)
print('}', file=ofile)
print('puts stdout "Done."', file=ofile)
print('quit -noprompt', file=ofile)
# Run magic to read in the individual GDS files and
# write out the consolidated GDS library
print('Running magic to create GDS library.')
sys.stdout.flush()
subprocess_run(
'magic',
['magic', '-dnull', '-noconsole',
destlibdir + '/generate_magic.tcl'],
cwd = destlibdir,
)
# Remove intermediate extraction files
extfiles = glob.glob(destlibdir + '/*.ext')
for extfile in extfiles:
os.remove(extfile)
# If the GDS file was a consolidated file of all cells, then
# create a similar SPICE library of all cells.
if os.path.isfile(allgdslibname):
spiext = '.spice' if not ef_format else '.spi'
create_spice_library(destlibdir, destlib, spiext, do_compile_only, do_stub, excludelist)
# Remove any files/directories that were marked for removal
for targname in removelist:
if os.path.isfile(targname):
os.remove(targname)
elif os.path.isdir(targname):
shutil.rmtree(targname)
sys.exit(0)