OpenTitan Earl Grey Chip Specification
This document describes the OpenTitan Earl Grey chip functionality in detail. For an overview, refer to the OpenTitan Earl Grey Chip Datasheet.
Theory of Operations
chip_earlgrey_asic contains the features listed above and is intended for ASIC synthesis, whereas the netlist
chip_earlgrey_cw310 provides an emulation environment for the
cw310 FPGA board.
The code for Ibex is developed in its own lowRISC repo, and is vendored in to this repository.
Surrounding Ibex is a suite of Comportable peripherals that follow the Comportability Guidelines for lowRISC peripheral IP.
Each of these IP has its own specification.
See the table produced in the hardware documentation page for links to those specifications.
This section provides some details for the processor and the peripherals. See their representative specifications for more information. This section also contains a brief overview of some of the features of the final product.
Clocking and Reset
Clocks and resets are supplied from the Analog Sensor Top, referred to as ast) from this point onwards in the document.
ast supplies a number of clocks into
- sys: main jittery system clock used for higher performance blocks and security (processory, memory and crypto blocks).
- io: a fixed clock used for peripheral blocks such as timers and I/O functionality, such as SPI or I2C.
- usb: a fixed clock used specifically for usb operations.
- aon: an always on, low frequency clock used for power management and low speed timers.
These clocks are then divided down and distributed to the rest of the system. See clock manager) for more details.
ast also supplies a number of power-okay signals to
top_earlgrey, and these are used as asynchronous root resets.
- vcaon_pok: The always on domain of the system is ready.
- vcmain_pok: The main operating domain of the system is ready.
When one of these power-okay signals drop, the corresponding domain in
top_earlgrey is reset.
Please refer to reset manager) for more details.
Resets throughout the design are asynchronous active low as per the Comportability specification.
Once reset, the reset vector begins in ROM, whose job is to validate code in the embedded flash before jumping to it. Valid code is assumed to have been instantiated into the flash, if not, the ROM shuts down the device unless prompted to bootstrap.
There are multiple avenues to load valid code into the flash:
- JTAG initiated flash programming.
- ROM bootstrap
AST Clocking and Reset Relationship
While the ast supplies clocks and resets to the
top_earlgrey, it also contains additional functions that interact with the design.
These include the
ADC, jittery clock controls and an assortment of other sensors.
The operating clocks and resets for these interfaces are supplied by the device in order to ensure correct synchronous operations.
The clock mapping is shown below:
The reset clock domain is identical to the table above, and the power domain mapping is shown below
System Reset Handling and Flash
top_earlgrey contains flash, it is important to examine the memory’s relationship with resets.
For flash, resets that occur during a stateful operation (program or erase) must be carefully handled to ensure the flash memory is not damaged. There are three reset scenarios:
- Reset due to external supply lowering.
- Reset due to internal peripheral request.
- Reset due to lower power entry and exit.
Reset due to External Supply
Device resets due to supply dropping below a specific threshold are commonly known as “brown-out”. When this occurs, the flash memory must go through specialized sequencing to ensure the cells are not damaged. This process is handled exclusively between ast and the flash. Please see the relevant section for more details.
Reset due to Internal Request
When the device receives an internal request to reset (for example aon_timer), device power is kept on and the flash is directly reset. It is assumed that the flash device, when powered, will be able to correctly handle such a sequence and properly protect itself.
Reset due to Low Power Entry
When the device receives a low power entry request while flash activity is ongoing, the pwrmgr) is responsible for ensuring the entry request is aborted.
Main processor (
The main processor (
core_ibex) is a small and efficient, 32-bit, in-order RISC-V core with a 2-stage pipeline that implements the RV32IMC instruction set architecture.
It was initially developed as part of the PULP platform under the name “Zero-riscy” , and has been contributed to lowRISC who maintains it and develops it further.
See the core_ibex specification for more details of the core.
In addition to the standard RISC-V functionality, Ibex implements M (machine) and U (user) mode per the RISC-V standard.
Attached to the Ibex core are a debug module (DM) and interrupt module (PLIC).
JTAG / Debug module
One feature available for Earl Grey processor core is debug access. By interfacing with JTAG pins, logic in the debug module allows the core to enter debug mode (per RISC-V 0.13 debug spec), and gives the design the ability to inject code either into the device - by emulating an instruction - or into memory. Full details can be found in the rv_dm specification.
Adjacent to the Ibex core is an interrupt controller that implements the RISC-V PLIC standard. This accepts a vector of interrupt sources within the device, and assigns leveling and priority to them before sending to the core for handling. See the details in the rv_plic specification.
Ibex currently achieves a CoreMark per MHz of 2.36 on the earlgrey verilator system. Performance improvements are ongoing, including the following items being considered:
- Adding a new ALU to calculate branch targets to remove a cycle of latency on taken conditional branches (currently the single ALU is used to compute the branch condition then the branch target the cycle following if the branch is taken).
- A 3rd pipeline stage to perform register writeback, this will remove a cycle of latency from all loads and stores and prevent a pipeline stall where a response to a load or store is available the cycle after the request.
- Implement a single-cycle multiplier.
- Produce an imprecise exception on an error response to a store allowing Ibex to continue executing past a store without waiting for the response.
The method for including these features, e.g. whether they will be configurable options or not, is still being discussed.
The Ibex documentation has more details on the current pipeline operation, including stall behaviour for each instruction in the Pipeline Details section.
The CoreMark performance achieved relies in part on single-cycle access to instruction memory. An instruction cache is planned to help maintain this performance when using flash memory that will likely not have single-cycle access times.
CoreMark was compiled with GCC 9.2.0 with flags:
-march=rv32imc -mabi=ilp32 -mcmodel=medany -mtune=sifive-3-series -O3 -falign-functions=16 -funroll-all-loops -finline-functions -falign-jumps=4 -mstrict-align
The device contains three memory address spaces for instruction and data.
Instruction ROM (32kB) is the target for the Ibex processor after release of external reset. The ROM contains hard-coded instructions whose purpose is to do a minimal subset of platform checking before checking the next stage of code. The next stage - a boot loader stored in embedded flash memory - is the first piece of code that is not hard-coded into the silicon of the device, and thus must be signature checked. The ROM executes this signature check by implementing a RSA-check algorithm on the full contents of the boot loader. The details of this check will come at a later date. For verification execute-time reasons, this RSA check will be overridable in the FPGA and verification platforms (details TBD). This is part of the Secure Boot Process that will be detailed in a security section in the future.
Earl Grey contains 1024kB of embedded-flash (e-flash) memory for code storage. This is intended to house the boot loader mentioned above, as well as the operating system and application that layers on top. At this time there is no operating system provided; applications are simple proof of concept code to show that the chip can do with a bare-metal framework.
Embedded-flash is the intended technology for a silicon design implementing the full OpenTitan device. It has interesting and challenging parameters that are unique to the technology that the silicon is implemented in. Earl Grey, as an FPGA proof of concept, will model these parameters in its emulation of the memory in order to prepare for the replacement with the silicon flash macros that will come. This includes the read-speeds, the page-sized erase and program interfaces, the two-bank update scheme, and the non-volatile nature of the memory. Since by definition these details can’t be finalized until a silicon technology node is chosen, these can only be emulated in the FPGA environment. We will choose parameters that are considered roughly equivalent of the state of the art embedded-flash macros on the market today.
Details on how e-flash memory is used by software will be detailed in future Secure Boot Process and Software sections over time.
The intent is for the contents of the embedded flash code to survive FPGA reset as it would as a NVM in silicon. Loading of the FPGA with initial content, or updating with new content, is described in other software specifications. The SPI device peripheral is provided as a method to bulk-load e-flash memory. The processor debug port (via JTAG) is also available for code loading. See those specifications for more details.
Also included is a 128kB of SRAM available for data storage (stack, heap, etc.) by the Ibex processor. It is also available for code storage, though that is not its intended purpose.
The base address of the ROM, Flash, and SRAM are given in the address map section later in this document.
Earl Grey contains a suite of “peripherals”, or subservient execution units connected to the Ibex processor by means of a bus interconnect. Each of these peripherals follows an interface scheme dictated in the Comportability Specification.. That specification details how the processor communicates with the peripheral (via TLUL interconnect); how the peripheral communicates with the chip IO (via fixed or multiplexable IO); how the peripheral communicates with the processor (interrupts); and how the peripheral communicates security events (via alerts). See that specification for generic details on this scheme.
Chip IO Peripherals
Pin Multiplexor Module (
The pin multiplexor’s purpose is to route between peripherals and the available multiplexable IO of the chip. At this time, the pin multiplexor is provided, but it is not used to its full potential. In addition, the multiplexor device manages control or pad attributes like drive strength, technology (OD, OS, etc), pull up, pull down, etc., of the chip’s external IO. It is notable that there are many differences between an FPGA implementation of Earl Grey and an ASIC version when it comes to pins and pads. Some pad attributes with analog characteristics like drive strength, slew rate and Open Drain technology are not supported on all platforms.
The pin multiplexor is a peripheral on the TLUL bus, with collections of registers that provide software configurability. See the pinmux specification for how to connect peripheral IO to chip IO and for information on pad control features.
The chip contains one UART peripheral that implement single-lane duplex UART functionality. The outputs and inputs can be configured to any chip IO via the pinmux.
See the UART specification for more details on this peripheral.
The chip contains one GPIO peripheral that creates 32 bits of bidirectional communication with the outside world via the pinmux. Via pinmux any of the 32 pins of GPIO can be connected to any of the 32 MIO chip pins, in any direction. See the GPIO specification for more details on this peripheral. See the pinmux specification for how to connect peripheral IO to chip IO.
The SPI device implements multiple modes:
- Firmware mode
- TPM mode
- Flash mode
- Passthrough mode
Firmware Mode is a generic data transfer mode. It can be used by software to construct a simple firmware upgrade mechanism for in-field devices.
TPM mode supports SPI transfers in compliance with the TPM PC Client Platform. The interface is single data lane and supports built-in data back pressuring.
Flash mode supports serial flash emulation. Typical commands such as read status, address mode, read data and program data are either natively supported or can be emulated.
Passthrough mode supports serial flash passthrough from an upstream SPI host to a downstream serial flash device.
See the SPI device specification for more details.
The chip contains a single module that supports USB device mode at full speed. In addition to normal functionality, USB suspend / resume is supported alongside the chip’s low power modes.
See the USB device specification for more details.
In order to be able to command I2C devices on systems where Earl Grey will be included, I2C host functionality will be required. This will include standard, full, and fast mode, up to 1Mbaud. More details of the I2C host module will come in a later specification update. The pins of the I2C host will be available to connect to any of the multiplexable IO (MIO) of the Earl Grey device. More than one I2C host module might be instantiated in the top level.
AES is the primary symmetric encryption and decryption mechanism used in OpenTitan protocols. AES runs with key sizes of 128b, 192b, or 256b. The module can select encryption or decryption of data that arrives in 16 byte quantities to be encrypted or decrypted using different block cipher modes of operation. It supports ECB mode, CBC mode, CFB mode, OFB mode and CTR mode.
The GCM mode is not implemented in hardware, but can be constructed using AES in counter mode. The integrity tag calculation can be implemented in Ibex and accelerated via bitmanip instructions.
Details on how to write key and data material into the peripheral, how to initiate encryption and decryption, and how to read out results, are available in the AES specification.
SHA-256 is the primary hashing algorithm used in OpenTitan protocols. SHA-256 is a member of the SHA-2 family of hashing algorithms, where the digest (or hash output) is of 256b length, regardless of the data size of the input to be hashed. The data is sent into the SHA peripheral after declaring the beginning of a hash request (effectively zeroing out the internal state to initial conditions), 32b at a time. Once all data has been sent, the user can indicate the completion of the hash request (with optional partial-word final write). The peripheral produces the hash result available for register read by the user. All data transfer is processor-available, i.e. data is passed into the module via register writes.
HMAC is a message authentication protocol layered on top of a hashing function (in this case SHA-256), mixing in a secret key for cryptographic purposes. HMAC is a particular application of appending the secret key in a prescribed manner, twice, around the hashing (via SHA-256) of the message.
Details on how to write key and data material into the peripheral, how to initiate hashing / authentication, and how to read out results, are available in the SHA/HMAC specification.
Alerts, as defined in the Comportability Specification, are defined as security-sensitive interrupts that need to be handled in a timely manner to respond to a security threat. Unlike standard interrupts, they are not solely handled by software. Alerts trigger a first-stage request to be handled by software in the standard mode as interrupts, but trigger a second-stage response by the alert handler if software is not able to respond. This ensures that the underlying concern is guaranteed to be addressed if the processor is busy, wedged, or itself under attack.
Each peripheral has an option to present a list of individual alerts, representing individual threats that require handling. These alerts are sent in a particular encoding method to the alert handler module, itself a peripheral on the system bus. See the details of the alert handler specification for more information.
TRNG entropy source
Randomness is a critical part of any security chip. It provides variations in execution that can keep attackers from predicting when the best time is to attack. It provides secret material used for identity and cryptographic purposes. It can be seeded into algorithmic computation to obscure sensitive data values. In short, it is a source of critical functionality that must be designed to be truly random, but also free from attack itself.
Most TRNGs (True Random Number Generators) are analog designs, taking advantage of some physical event or process that is non-deterministic. Example designs rely on metastability, electronic noise, timing variations, thermal noise, quantum variation, etc. These are then filtered and sent into a pool of entropy that the device can sample at any time, for whatever purposes are needed. The creation, filtering, storage, protection, and dissemination of the randomness are all deep topics of intense research in their own right.
The primary interface to the entropy pool is a read request of available random bits. The TRNG interface can indicate how many bits are available, and then software can read from this pool, if available. Reading of entropy that is not available should immediately trigger an interrupt or an alert.
Since silicon is required to contain an analog design tied to the final chosen silicon technology process, our FPGA implementation can only approximate the results. We however fully specify the software interface to the TRNG in a digital wrapper. In FPGA we emulate the randomness with something akin to a PRBS.
Timers are critical for operating systems to ensure guaranteed performance for users. To some level they are even required by the RISC-V specification. At this time, one timer is provided, a 64b free running timer with a guaranteed (within a certain percentage) frequency. A second one acting as a watchdog timer that can be used to backstop the processor in the case of it being unresponsive (usually due to development code that is wedged, rather than for instance due to security attack) will be provided in the future. The goal is for both of these to be satisfied with the same timer module.
The specification for the timer can be found here.
The final peripheral discussed in this release of the netlist is an emulated flash controller. As mentioned in the memory section, up to 1024kB of emulated embedded flash is available for code and data storage. The primary read path for this data is in the standard memory address space. Writes to that address space are ignored, however, since one can not write to flash in a standard way. Instead, to write to flash, software must interact with the flash controller.
Flash functionality include three primary commands: read, erase, and program.
Read, as mentioned above, is standard, and uses the chip memory address space.
Erase is done at a page level, where the page size is parameterizable in the flash controller.
Upon receiving an erase request, the flash controller wipes all contents of that page, rendering the data in all
1s state (
0xFFFFFFFF per word).
Afterwards, software can program individual words to any value.
It is notable that software can continue to attempt to program words even before another erase, but it is not physically possible to return a flash bit back to a
'1' state without another erase.
So future content is in effect an AND of the current content and the written value.
next_value = AND(current_value, write_word)
Erase and program are slow. A typical erase time is measured in milliseconds, program times in microseconds. The flash controller peripheral in this release approximates those expected times.
Security is also a concern, since secret data can be stored in the flash. Some memory protection is provided by the flash controller. For more details see the flash controller module specification.
Interconnecting the processor and peripheral and memory units is a bus network built upon the TileLink-Uncached-Light protocol. See the OpenTitan bus specification for more details.
top_earlgrey has a two-level hierarchy for its bus network.
Close to the CPU is the high-speed cluster, with low access latency.
Farther from the CPU, through a low-speed bridge, is the low-speed cluster, with higher access latency.
The following table describes the typical number of CPU cycles from a transaction’s launch to its completion for a device in the specified cluster. Note that these values assume there is no bus contention.
|Cluster||CPU Access Latency|
|High speed||2 CPU cycles|
|Low speed||18-20 CPU cycles|
The base addresses of the memory and peripherals are given in the table below.
The choice of memory, or lack thereof at location 0x0 confers two exclusive benefits:
- If there are no memories at location 0x0, then null pointers will immediately error and be noticed by software (the xbar will fail to decode and route)
- If SRAM is placed at 0, accesses to data located within 2KB of 0x0 can be accomplished with a single instruction and thus reduce code size.
For the purpose of
top_earlgrey, the first option has been chosen to benefit software development and testing
|1||USB_P||VCC||DualBidirTol||manual||USB P signal|
|2||USB_N||VCC||DualBidirTol||manual||USB N signal|
|3||CC1||AVCC||InputStd||manual||ADC input 1|
|4||CC2||AVCC||InputStd||manual||ADC input 2|
|5||FLASH_TEST_VOLT||VCC||AnalogIn0||manual||Flash test voltage input|
|6||FLASH_TEST_MODE0||VCC||InputStd||manual||Flash test mode signal|
|7||FLASH_TEST_MODE1||VCC||InputStd||manual||Flash test mode signal|
|8||OTP_EXT_VOLT||VCC||AnalogIn1||manual||OTP external voltage input|
|9||SPI_HOST_D0||VIOA||BidirStd||direct||SPI host data|
|10||SPI_HOST_D1||VIOA||BidirStd||direct||SPI host data|
|11||SPI_HOST_D2||VIOA||BidirStd||direct||SPI host data|
|12||SPI_HOST_D3||VIOA||BidirStd||direct||SPI host data|
|13||SPI_HOST_CLK||VIOA||BidirStd||direct||SPI host clock|
|14||SPI_HOST_CS_L||VIOA||BidirStd||direct||SPI host chip select|
|15||SPI_DEV_D0||VIOA||BidirStd||direct||SPI device data|
|16||SPI_DEV_D1||VIOA||BidirStd||direct||SPI device data|
|17||SPI_DEV_D2||VIOA||BidirStd||direct||SPI device data|
|18||SPI_DEV_D3||VIOA||BidirStd||direct||SPI device data|
|19||SPI_DEV_CLK||VIOA||InputStd||direct||SPI device clock|
|20||SPI_DEV_CS_L||VIOA||InputStd||direct||SPI device chip select|
|0||IOA0||VIOA||BidirStd||muxed||Muxed IO pad|
|1||IOA1||VIOA||BidirStd||muxed||Muxed IO pad|
|2||IOA2||VIOA||BidirStd||muxed||Muxed IO pad|
|3||IOA3||VIOA||BidirStd||muxed||Muxed IO pad|
|4||IOA4||VIOA||BidirStd||muxed||Muxed IO pad|
|5||IOA5||VIOA||BidirStd||muxed||Muxed IO pad|
|6||IOA6||VIOA||BidirOd||muxed||Muxed IO pad|
|7||IOA7||VIOA||BidirOd||muxed||Muxed IO pad|
|8||IOA8||VIOA||BidirOd||muxed||Muxed IO pad|
|9||IOB0||VIOB||BidirStd||muxed||Muxed IO pad|
|10||IOB1||VIOB||BidirStd||muxed||Muxed IO pad|
|11||IOB2||VIOB||BidirStd||muxed||Muxed IO pad|
|12||IOB3||VIOB||BidirStd||muxed||Muxed IO pad|
|13||IOB4||VIOB||BidirStd||muxed||Muxed IO pad|
|14||IOB5||VIOB||BidirStd||muxed||Muxed IO pad|
|15||IOB6||VIOB||BidirStd||muxed||Muxed IO pad|
|16||IOB7||VIOB||BidirStd||muxed||Muxed IO pad|
|17||IOB8||VIOB||BidirStd||muxed||Muxed IO pad|
|18||IOB9||VIOB||BidirOd||muxed||Muxed IO pad|
|19||IOB10||VIOB||BidirOd||muxed||Muxed IO pad|
|20||IOB11||VIOB||BidirOd||muxed||Muxed IO pad|
|21||IOB12||VIOB||BidirOd||muxed||Muxed IO pad|
|22||IOC0||VCC||BidirStd||muxed||Muxed IO pad|
|23||IOC1||VCC||BidirStd||muxed||Muxed IO pad|
|24||IOC2||VCC||BidirStd||muxed||Muxed IO pad|
|25||IOC3||VCC||BidirStd||muxed||Muxed IO pad|
|26||IOC4||VCC||BidirStd||muxed||Muxed IO pad|
|27||IOC5||VCC||BidirStd||muxed||Muxed IO pad|
|28||IOC6||VCC||BidirStd||muxed||Muxed IO pad|
|29||IOC7||VCC||BidirStd||muxed||Muxed IO pad|
|30||IOC8||VCC||BidirStd||muxed||Muxed IO pad|
|31||IOC9||VCC||BidirStd||muxed||Muxed IO pad|
|32||IOC10||VCC||BidirOd||muxed||Muxed IO pad|
|33||IOC11||VCC||BidirOd||muxed||Muxed IO pad|
|34||IOC12||VCC||BidirOd||muxed||Muxed IO pad|
|35||IOR0||VCC||BidirStd||muxed||Muxed IO pad|
|36||IOR1||VCC||BidirStd||muxed||Muxed IO pad|
|37||IOR2||VCC||BidirStd||muxed||Muxed IO pad|
|38||IOR3||VCC||BidirStd||muxed||Muxed IO pad|
|39||IOR4||VCC||BidirStd||muxed||Muxed IO pad|
|40||IOR5||VCC||BidirStd||muxed||Muxed IO pad|
|41||IOR6||VCC||BidirStd||muxed||Muxed IO pad|
|42||IOR7||VCC||BidirStd||muxed||Muxed IO pad|
|21||IOR8||VCC||BidirOd||direct||Dedicated sysrst_ctrl output (ec_rst_l)|
|22||IOR9||VCC||BidirOd||direct||Dedicated sysrst_ctrl output (flash_wp_l))|
|43||IOR10||VCC||BidirOd||muxed||Muxed IO pad|
|44||IOR11||VCC||BidirOd||muxed||Muxed IO pad|
|45||IOR12||VCC||BidirOd||muxed||Muxed IO pad|
|46||IOR13||VCC||BidirOd||muxed||Muxed IO pad|
RTL Implementation Notes
At this time, the top-level netlist for earlgrey is a combination of hand-written SystemVerilog RTL with auto-generated sections for wiring of comportability interfaces.
There is a script for this auto-generation, centered around the top-level descriptor file
top_earlgrey.hjson found in the repository.
A full definition of this descriptor file, its features, and related scripting is forthcoming.
This tooling generates the interconnecting crossbar (via
TLUL) as well as the instantiations at the top level.
It also feeds into this document generation to ensure that the chosen address locations are documented automatically using the data in the source files.
Top Level vs Chip Targets
It should first be NOTED that there is some subtlety on the notion of hierarchy within the top level.
There is netlist automation to create the module
top_earlgrey as indicated in sections of this specification that follow.
On top of that module, hierarchically in the repo, are chip level instantiation targets directed towards a particular use case.
chip_earlgrey_cw310 for use in FPGA, and
chip_earlgrey_asic for use (eventually) in a silicon implementation.
These chip level targets will include the actual pads as needed by the target platform.
At the time of this writing the two are not in perfect synchronization, but the intention will be for them to be as identical as possible.
Where appropriate, including the block diagram below, notes will be provided where the hierarchy subtleties are explained.
TODO: This section needs to be updated once pin updates are complete.
In the FPGA platform, the logic for the JTAG and the SPI device are multiplexed within
This is done for ease of programming by the external host.