Main

Recent years have witnessed a surge in photodetector research based on emerging semiconducting materials (for example, organic semiconductors, halide perovskites and derivatives, quantum dots, two-dimensional semiconductors, metal oxides and carbon nanotubes1,2,3,4,5,6,7,8,9,10,11), with a focus on devices relying on photocarrier generation and extraction within the device stack (internal photoelectric effect; Supplementary Notes 1 and 2).

Technological advances are tied to accurate performance characterization and benchmarking. However, the literature on emerging photodetector technologies frequently exhibits inconsistencies in characterization and reporting, posing challenges for benchmarking. These inconsistencies may partly stem from the diverse fields involved in developing these technologies, as well as the emergence of new and distinctive effects, creating difficulties in establishing a common language. Consistency issues also arise from the diversity of target applications, technologies and spectral ranges, as well as the knowledge gap between academic and industrial research. Academic efforts have often focused on optimizing a narrow set of performance metrics, which may not always align with the broader application needs pursued by industry. Although a few publications have offered valuable insights into some of these issues12,13,14,15,16, a comprehensive examination covering the characterization, reporting and benchmarking of all key performance parameters for emerging photodetector technologies (alongside detailed guidelines) is currently missing.

Since these challenges impact a vast research community and, therefore, require collective discussion, we have come together as a team representing diverse streams in emerging photodetector technologies, spanning both academia and industry. On the basis of a collaborative effort, we present a Consensus Statement providing comprehensive guidelines for accurate performance characterization, reporting and benchmarking for emerging photodetector technologies. Detailed tabulated guidelines are provided in Supplementary Tables 113.

Dark current

Current–voltage sweeps are typically used to characterize the dark current (Idark) in emerging photodetector technologies (Supplementary Notes 3 and 4). However, this method is unsuitable if transient effects occur during current–voltage sweeps (Fig. 1a; for example, due to charge trapping, capacitive charging/discharging or ion/defect migration)16,17. Although minimizing these effects is essential for most application scenarios and requires suitable mitigation strategies16,18, such strategies may not be fully developed for early stage technologies that otherwise show promise. In cases where such transient effects are present and mitigation strategies are still under development, accurate Idark measurements require sufficient time for the current to stabilize after applying a voltage19,20 (Fig. 1b). Moreover, reporting the transient behaviour of the dark current on step-voltage application (Fig. 1b) is valuable for optimizing photodetectors targeting imaging applications (Supplementary Note 5).

Fig. 1: Dark-current diagnostics.
figure 1

a, Double-sweep current–voltage curves that may be observed from a photodiode in the dark, revealing sweep-rate dependence and hysteresis. b, Transient dark-current effects that may occur following step-voltage application.

Early stage emerging photodetector technologies may also display Idark drift during the measurement of various other performance parameters. In such cases, monitoring the evolution of Idark over extended periods is crucial. For reliable benchmarking, we advise adopting sufficiently long stabilization periods to ensure that variations in Idark remain below 5% from the beginning to the end of a spectral responsivity measurement (Supplementary Note 6). Accurate reporting should include the time evolution of Idark over this timescale (with the photodetector biased as in a spectral responsivity measurement), with appropriate additional checks to confirm the minimal presence of longer-term transients.

In analysing device performance, it is often necessary to determine the areal dark-current density (jdark), especially in devices with charge transport in the out-of-plane direction (Supplementary Note 7). A simple normalization of the measured Idark by the nominal active area is often inaccurate for estimating jdark for emerging photodetector technologies13. This is because they often feature semiconducting and/or charge transport layers extending considerably beyond the interelectrode region, leading to dark-current contributions from outside this area, alongside a spatially non-uniform jdark. Therefore, for accurately determining jdark, we recommend fabricating photodetectors with varying electrode and active area sizes to identify the appropriate area for normalizing Idark.

As the Idark of emerging photodetector technologies may be strongly influenced by temperature, it is important to report the temperature at which it is measured and the Idark values across the temperature range relevant to the target application to enable meaningful benchmarking.

Responsivity and external quantum efficiency

To accurately determine responsivity and external quantum efficiency (EQE; Supplementary Note 3), the incident optical power spectrum should be measured and consistently monitored with a calibrated reference detector (Supplementary Notes 8 and 9 and Supplementary Fig. 1). Moreover, the illuminated area must be strictly defined for both reference detector and photodetector under test. An ideal solution is to use an aperture that covers 95%–100% of the active area of the photodetector under test, ensuring that the irradiance within this area remains spatially uniform within 5% of its maximum (Fig. 2a, Supplementary Fig. 2 and Supplementary Notes 8 and 10). Overfilling the active area without using such an aperture may cause mischaracterization, as photons absorbed outside the active area can contribute to the photocurrent21 (Fig. 2a). Moreover, although responsivity and EQE are generally properties of a photodetector as a whole, underfilling the active area assesses only the response of the illuminated portion, potentially misrepresenting the device performance (particularly for emerging photodetector technologies that often exhibit spatial non-uniformities). This problem is exacerbated when the underfilling beam reaches irradiance levels outside the photodetector’s linear dynamic range (LDR) in some regions of its cross-section. This leads to spatially varying responsivity and EQE due to non-uniform illumination (Supplementary Note 8). In general, illumination should be collimated (or pseudo-collimated) and normal to the photodetector plane (unless the angular response is being measured) to avoid inaccuracies due to the angular interplay among the light source, reference detector and photodetector under test22.

Fig. 2: Challenges in responsivity/EQE characterization.
figure 2

a, Left: overfilling incident beam causing inaccurate responsivity/EQE determination due to carrier generation beyond the nominal active area, extending into a peripheral region of feature size DL. Right: aperture adoption for accurate responsivity/EQE determination. b, Importance of monitoring the time-dependent dark current for accurate responsivity/EQE measurements, during which the device is subjected to multiple optical pulses with power Pi(t). c, Impact of unsuitable modulation frequency on modulated measurements for continuous-wave responsivity/EQE characterization. d, Time-domain illustration of the effect displayed in c. Top: suitable modulation frequency, allowing the accurate determination of the continuous-wave responsivity/EQE from a modulated measurement. Bottom: unsuitable modulation frequency.

For devices with instability or appreciable transient effects, monitoring Idark throughout a responsivity/EQE measurement is necessary to accurately determine the photocurrent (Fig. 2b). Although modulated responsivity/EQE measurements purge the current recorded under the illumination of contributions from a slowly varying background, they are nonetheless affected by drifts in device characteristics. Therefore, for devices with instability or appreciable transient effects, allowing them to stabilize before a responsivity/EQE measurement is recommended23,24,25. Moreover, in modulated measurements to determine the continuous-wave responsivity/EQE, the modulation frequency should be much smaller than the photodetector bandwidth (Fig. 2c,d), with responsivity/EQE varying ≤5% across a frequency decade centred on the modulation frequency.

The spectral responsivity R(λ) (λ, wavelength) and EQE spectrum should be characterized over the entire spectral range of the incident optical signals relevant to the application under consideration (reference spectral range; Supplementary Note 3 and Supplementary Fig. 3), with the incident optical power at each wavelength consistently monitored with a calibrated reference detector. For benchmarking purposes, we recommend referring to the conventional reference spectral ranges presented in Supplementary Table 14.

In the literature covering spectrally selective photodetectors (Supplementary Note 3 and Supplementary Fig. 3), the full width at half maximum of the spectral responsivity or EQE spectrum (FWHMR or FWHMEQE) is sometimes incorrectly considered as a parameter that should necessarily be minimized to achieve high performance. In fact, the necessary FWHMR and FWHMEQE depend on the application of interest. Therefore, emphasis on their minimization should only be reserved to spectrally selective photodetectors targeting multispectral or hyperspectral light sensing. Furthermore, photodetectors lacking a clear responsivity or EQE passband (that is, having a responsivity/EQE that may not reach half maximum or exhibiting multiple passbands within the reference spectral range) are sometimes erroneously referred to as spectrally selective. For accurate reporting, only devices with a well-defined FWHMR or FWHMEQE should be presented as spectrally selective.

Photoconductive gain versus EQE

In reports on devices with injecting electrodes, the emerging photodetector literature sometimes incorrectly uses photoconductive gain (G) and EQE interchangeably (Supplementary Notes 3 and 11). In fact, the two are conceptually and quantitatively distinct: EQE measures the number of electrons collected at the contacts per incident photon per unit time, whereas photoconductive gain refers to the number of collected carriers (as a result of illumination) per photogenerated electron–hole pair (Supplementary Notes 3 and 12).

Furthermore, high apparent EQE (100%) and G (1) in devices with injecting electrodes do not necessarily indicate superior optoelectronic performance. In devices without photoconductive gain, a high EQE genuinely reflects excellent optoelectronic performance. However, devices with a large photoconductive gain often suffer from a slow response (potentially, persistent changes in conductivity)25, making the gain–bandwidth product a more appropriate benchmarking metric26,27,28. Moreover, gain can increase dark current and noise (neglecting this can lead to overestimated specific detectivity and dynamic range12,29). Additionally, gain typically depends on trap filling, which can introduce nonlinear irradiance dependence of the photoresponse30,31. Therefore, superlative claims based solely on high apparent EQE and G in such devices should be avoided. Instead, their evaluation should consider how gain holistically impacts device performance.

From a practical perspective, the emphasis on gain versus speed of response varies with application. For instance, fast response is critical for high-speed communication, whereas higher gain may be more important for imaging under low irradiance. Moreover, recent work has highlighted photodetectors with very slow recovery times for use in neuromorphic devices32. Therefore, application-specific contextualization is required when photoconductive gain is reported.

Linearity and LDR

A misunderstanding often encountered in the emerging photodetector literature is that straight light current (Io)–optical power (Pi) characteristics plotted on a double-logarithmic scale (Fig. 3a) imply linearity (Supplementary Note 3 and Supplementary Fig. 4). In fact, such straight characteristics with slopes α different from unity correspond to the power law \({I}_{o}\propto \,{P}_{i}^{\alpha }\), where α ≠ 1 implies nonlinearity (Fig. 3b). To prevent this misunderstanding, researchers should report the slope α of the log[Io]–log[Pi] dataset fit (Fig. 3b). The conclusion of linearity should only be drawn when α is practically indistinguishable from unity. In general, the acceptable deviation from unity depends on the application. For benchmarking purposes, however, we recommend referring to a photodetector as strictly linear if α falls within [0.99, 1.01] when rounded to the second decimal place, and as quasi-linear if α falls within [0.97, 0.99) and (1.01,1.03] (Supplementary Note 13 and Supplementary Fig. 5). Additionally, the responsivity–optical power plot should be provided for further validation (Fig. 3c).

Fig. 3: Linearity and LDR in photodetectors.
figure 3

a, Representative current–power curve for a linear photodetector and key linearity-related parameters. b, Comparison of current–power curves for linear (α= 1) and nonlinear (α ≠ 1) photodetectors. c, Responsivity versus optical power corresponding to b. d, Sketch of an experimental current–power curve (solid line) for a linear photodetector. Outside the measured range, the curve might evolve as linear or nonlinear (dotted traces).

A related common misconception is the quantification of the LDR for photodetectors with straight IoPi characteristics on a double-logarithmic scale but for which α ≠ 1. This is generally incorrect, as LDR is a concept exclusive to linear photodetectors (that is, α= 1).

In reporting the LDR of linear photodetectors, confusion surrounds the prefactor (20 versus 10) used in its definition, leading to inconsistent values across publications. The formula to be adopted is

$${\rm{LDR}}{{\stackrel{\scriptscriptstyle\mathrm{def}}{=}}}20{\log }_{10}\left[\frac{{I}_{{\rm{o}},{\rm{M}}}}{{I}_{{\rm{o}},{\rm{m}}}}\right].$$
(1)

Io,M and Io,m denote the maximum and minimum light current values, respectively, within which the photodetector behaves linearly (Fig. 3a). Arguments supporting a prefactor of 10 often stem from the expression of LDR in terms of optical power (that is, LDR  log10[Pi,M/Pi,m], where Pi,M and Pi,m are the optical power values corresponding to Io,M and Io,m, respectively). This association with a prefactor of 10 is drawn from its use in the definition of decibels for signal power (defined as a quantity quadratically dependent on the signal) in signal theory. However, this perspective is inaccurate because, in the context of LDR, optical power serves as a proxy for photon flux, which does not have a quadratic dependence on the signals at play (for example, the photocurrent and the number of incident photons).

To address the ambiguity often found in the literature concerning the determination of Io,M and Io,m to calculate the LDR, the tolerable deviation Δ of the measured dataset from linearity should be quantified:

$$\pm \varDelta {{\stackrel{\scriptscriptstyle\mathrm{def}}{=}}}\frac{\text{d}\log_{10} [{I}_{\text{o}}]}{\text{d}\log_{10} [{P}_{\text{i}}]}-1.$$
(2)

Although Δ is generally application dependent, Δ = 0.01 and Δ = 0.03 are recommended for benchmarking emerging photodetector technologies that are strictly linear and quasi-linear, respectively (Supplementary Note 13).

Io,m is sometimes erroneously equated to Idark, or the lowest light current measured with the equipment at hand, or the (calculated or measured) root-mean-square current noise in,r.m.s. (noise floor). Similarly, Io,M is often incorrectly equated to the maximum light current measured with the equipment at hand. These approaches are incorrect because LDR is defined with respect to the actual extrema of a photodetector’s linear range, instead of estimated values. Moreover, without experimental validation, it is incorrect to assume that the linear trend observed at higher optical powers extends to in,r.m.s. (Fig. 3d). If limitations of the characterization apparatus prevent reaching the actual extrema of the linear range, the apparent LDR (LDRapp) should be reported instead, referring to it as such (Fig. 3d). LDRapp extends between the apparent minimum and maximum measured light current values (Io,m,app and Io,M,app, respectively) over which linearity holds (that is, α within [0.99, 1.01] for strictly linear photodetectors or [0.97, 0.99) and (1.01,1.03] in the quasi-linear case):

$${{{{\rm{LDR}}}}}_{{{\rm{app}}}}{{\stackrel{\scriptscriptstyle\mathrm{def}}{=}}}20\,{\log }_{10}\left[\frac{{I}_{{\rm{o}},{\rm{M}},{\rm{app}}}}{{I}_{{\rm{o}},{\rm{m}},{\rm{app}}}}\right].$$
(3)

By definition, LDR ≥ LDRapp. For benchmarking, if Io,m,app or Io,M,app is determined by deviation from linearity, it should be calculated through equation (2) with Δ = 0.01 or Δ = 0.03.

Although linear photodetector operation with a wide LDR is essential for most applications, some emerging areas (for example, in-sensor computing and neuromorphic devices) may benefit from nonlinear photodetectors33,34,35. Therefore, nonlinear photodetector studies should contextualize their findings within the latter areas, benchmarking their devices exclusively against the relevant literature on nonlinear photodetectors.

Noise

To avoid misreporting the noise performance of emerging photodetector technologies, it is essential to quantify and correct for instrumentation noise as part of the characterization of their noise power spectral density, alongside reporting all relevant experimental quantities (Supplementary Notes 3 and 14 and Supplementary Fig. 6).

The commonly used relation \(\sqrt{S_{n}(f)\,B}\) for quantifying the the root-mean-square value of the intrinsic photodetector noise, in,r.m.s., over an equivalent noise bandwidth B (Supplementary Note 15) is correct only if the power spectral density of the intrinsic photodetector noise (Sn(f), where f is the frequency) is frequency independent (white noise) over the frequency range of interest. For instance, for pink noise (Sn(f)  1/fβ, with the constant β typically ranging from 0.8 to 1.5)36,37,38, this relationship does not generally apply (potentially leading to large errors; Fig. 4a) and only approximately holds when Bf (Supplementary Note 16). Therefore, quantifying in,r.m.s. as \(\sqrt{S_{n}(f)\,B}\) requires prior validation of the frequency independence of Sn or other relevant assumptions (for example, Bf for dominant pink noise).

Fig. 4: Noise, NEP and detectivity.
figure 4

a, Relative error in estimating \({i}_{n,{r.m.s.}}^{2}\) if the formula Sn(f) B is used instead of the exact integral (Supplementary Note 16) to quantify the power of a pink-noise source (Sn(f)  1/f). b, Power spectral density of noise superimposed on a modulated signal with decreasing amplitude, eventually becoming indistinguishable from noise. c, Corresponding signal-to-noise ratio and NEP determination. d, Representative power spectral density of superimposed white and pink-noise components. e, Corresponding root-mean-square noise integrated over a bandwidth B, normalized to \(\sqrt{B}\). These results show that normalizing the root-mean-square noise by \(\sqrt{B}\) is not meaningful in the pink-noise-dominated region. The trace labelled ‘approx.’ refers to the generally inaccurate approximation \({i}_{n,{r.m.s}}=\sqrt{S_{n}(f)\,B}\). f, Specific detectivity corresponding to the data in d and e, calculated as \({D}^{* }\stackrel{\scriptscriptstyle\mathrm{def}}{=}{D}\sqrt{{AB}}\) assuming unity responsivity and area, revealing unreliable D* values in the pink-noise-dominated region.

In general, in,r.m.s. also depends on device geometry. For instance, in vertical photodiodes with no photoconductive gain, \({i}_{n,{r.m.s.}}\propto \sqrt{A}\) (A, nominal in-plane device area) when shot noise dominates. In such cases, normalizing in,r.m.s. by \(\sqrt{A}\) is both useful and accurate for cross-device comparisons. However, the functional dependence of in,r.m.s. on geometric parameters is not universally of the form \(\sqrt{A}\) (Supplementary Note 17). Therefore, normalizing in,r.m.s. by \(\sqrt{A}\) does not generally yield a geometry-independent noise assessment. In other words, using such normalization to compare the intrinsic noise of photodetectors whose in,r.m.s. does not solely depend on device geometry via \(\sqrt{A}\) would lead to erroneous conclusions. For this reason, accurate comparisons of geometry-normalized photodetector noise can only be made between devices with the same geometry and dominant noise mechanisms. Normalizing noise performance by geometric parameters requires prior experimentation and modelling to establish the specific dependence of in,r.m.s. on those parameters (Supplementary Note 18).

Experiments on emerging photodetector technologies have revealed that their noise power spectral density cannot be accurately determined through model formulas for white-noise sources—for example, via the shot-noise model39,40,41. However, we acknowledge that many researchers developing emerging photodetector technologies may lack access to apparatus for characterizing noise power spectral density. In such cases, reporting lower-limit noise values based on white-noise models must be accompanied by an explicit acknowledgement of their theoretical nature, with appropriate labelling (for example, using in,r.m.s.,theor. to denote the theoretical in,r.m.s.) to distinguish them from the experimental data. Importantly, such theoretical values should not be used as the basis for superlative performance claims, nor should they be compared with experimentally determined noise values from other technologies.

When theoretical, lower-limit noise values based on white-noise models are determined, choosing the appropriate model is crucial. For instance, thermal noise is often omitted without verifying whether this approximation is correct. Moreover, to calculate shot noise in devices with photoconductive gain G, the basic shot-noise model \({i}_{n,{r.m.s},{theor}.}=\sqrt{2q{I}_{\rm{dark}}B}\) (q, elementary charge) is not applicable and this alternative equation must be used29,42,43.

$${i}_{n,{r.m.s.},{theor}.}=\sqrt{2q{I}_{\rm{dark}}{GFB}}$$
(4)

Here F is the Fano factor (F = 1 if charge transport is due to uncorrelated, independent events). Using \({i}_{n,{r.m.s},{theor}.}=\sqrt{2q{I}_{\rm{dark}}B}\) instead of equation (4) can lead to substantial noise miscalculations (thereby introducing substantial inaccuracies in derived quantities), given that typically G 1. To mitigate confusion regarding the choice of models for calculating theoretical white-noise limits, we recommend using the following formula.

$${i}_{n,{r.m.s},{theor}.}=\sqrt{2q{I}_{\rm{dark}}{GF}B+4{k}_{B}TB/{R}_{\rm{dark}}^{(\rm{PD})}}$$
(5)

Here \({R}_{\rm{dark}}^{(\rm{PD})}\) is the differential photodetector resistance (dVbias/dIdark) in the dark around the same operating voltage Vbias used for Idark and photoresponse measurements, T is the absolute temperature and kB is Boltzmann’s constant. We encourage reporting the theoretical white-noise baseline using equation (5)—explicitly presenting it as a theoretical limit—to facilitate cross-study comparisons of theoretical noise performance. However, equation (5) should not be considered a substitute for experimental noise measurements, which remain essential for accurately evaluating noise performance.

Noise-equivalent power and detectivity

To quantify the noise-equivalent power (NEP) and detectivity (D; Supplementary Note 3 and Supplementary Fig. 7), we recommend acquiring the photodetector output power spectral density (purging it from instrument noise; Supplementary Note 14) as a modulated optical signal of root-mean-square optical power Pi,r.m.s. is applied, centred at frequency f and spanning a bandwidth B (Fig. 4b,c). The Pi,r.m.s. yielding a photodetector output signal equal to the background noise power spectral density quantifies the photodetector’s NEP at that frequency and bandwidth, whereas its reciprocal gives D.

Alternatively, if suitable conditions are verified, NEP and D can be determined from the ratio between the measured in,r.m.s. (for the selected frequency and bandwidth, which must be specified) and the measured responsivity (Supplementary Note 3). This approach is applicable only if the responsivity is measured at an optical power approaching NEP, which requires experimental validation and reporting. If linearity is assessed only at optical powers far above the NEP due to apparatus limitations, it is uncertain whether the apparent responsivity Rapp measured within the apparent LDR can be used to determine NEP and D. In such cases, the ratio Rapp/in,r.m.s. should be referred to as the apparent detectivity and labelled Dapp. Correspondingly, the apparent NEP, NEPapp, is calculated as 1/Dapp. Importantly, if a photodetector is nonlinear as Pi,r.m.s. approaches NEP, its responsivity varies with optical power in that region; hence, using the responsivity measured at an arbitrary optical power far above the NEP would lead to inaccurate NEP and D determination. In such cases, neither Dapp nor NEPapp is defined, and direct experimental evaluation (Fig. 4b,c) is the only viable approach to determining NEP and D.

For geometry- and bandwidth-independent benchmarking, the emerging photodetector literature commonly multiplies detectivity by the square root of the product of the nominal in-plane device area (A) and noise bandwidth (B), which defines the specific detectivity \({D}^{* }\stackrel{\scriptscriptstyle\mathrm{def}}{=}{D}\sqrt{{AB}}\). However, this normalization is meaningful only if in,r.m.s. depends on bandwidth and geometry solely through the term \(\sqrt{{AB}}\) (for example, as in vertical photodiodes with dominant shot noise). Therefore, D* should be reported only after experimental validation that in,r.m.s. exhibits this specific functional dependence on bandwidth and geometry (Supplementary Note 18).

If in,r.m.s. deviates from the \(\sqrt{{AB}}\) dependence, adopting D* as a benchmarking metric is misleading, as D* itself depends on geometry and noise bandwidth in such cases. For instance, in reference to the photodetector with representative noise power spectral density in Fig. 4d, if one laboratory were to measure noise with B = 1 Hz and another with B = 100 Hz, the latter may mistakenly conclude that they have measured a superior photodetector because of the higher D* (Fig. 4e,f). To overcome the limitations of the definition of D* as \({D}\sqrt{{AB}}\) for cases with a non-square-root bandwidth dependence of in,r.m.s. (but confirmed dependence on device geometry via \(\sqrt{A}\)), we recommend using a modified specific detectivity, \({D}_{\bar{B}}^{* }\), referenced to a fixed bandwidth \(\bar{B}\).

$${D}_{\bar{B}}^{* }\left(f,\lambda \right){{\stackrel{\scriptscriptstyle{\mathrm{def}}}{=}}}R\left(f,\lambda {,P}_{i,{r.m.s.}}={\rm{NEP}}\right)\sqrt{A}/{i}_{n,{r.m.s.}}(f,\bar{B})$$
(6)

Here R(f, λ, Pi,r.m.s. = NEP) is the responsivity measured at modulation frequency f, wavelength λ and Pi,r.m.s. = NEP. \(\bar{B}=1\) Hz is a convenient choice for benchmarking purposes, especially for applications operating at frequencies above 10 Hz.

In general, for a bandwidth- and geometry-independent comparison of detection capabilities across photodetectors for which in,r.m.s. is not proportional to \(\sqrt{{AB}}\), a different normalization would be necessary. In such cases, normalization should incorporate the specific functional dependencies of in,r.m.s. on bandwidth and geometry for the devices at hand. For the same reason, when dealing with photodetectors with distinct functional dependencies on geometry and bandwidth, it is advisable to consider detectivity (or NEP) rather than specific detectivity for a meaningful comparison of their detection limits. Indeed, in such cases, D* and \({D}_{\bar{B}}^{* }\) could be easily manipulated by varying device dimensions or bandwidth.

Finally, a widespread feature of specific detectivity reports on emerging photodetector technologies is the reliance on noise values obtained from theoretical noise models (for example, shot and thermal noise equations) instead of experimental noise measurements. This approach may be inaccurate up to several orders of magnitude, especially if pink noise dominates. If an apparatus for characterizing the power spectral density of photodetector noise is not readily available, resorting to such models is acceptable only if the resultant detectivity values are presented as theoretical estimates and labelled accordingly. It remains crucial, however, to base estimates on correct model equations. For instance, in devices with photoconductive gain, neglecting the impact of gain on shot noise results in exaggerated detectivity values12. Therefore, provided that the assumptions underlying equation (6) hold, it is useful, for benchmarking purposes, to report the theoretical specific detectivity in the white-noise limit, \({D}_{{theor}.}^{* }\left(f,\lambda \right)\), using the following equation.

$${D}_{{theor}.}^{* }\left(f,\lambda \right){{\stackrel{\scriptscriptstyle{\mathrm{def}}}{=}}}R\left(f,\lambda {,P}_{i,{r.m.s.}}={\rm{NEP}}\right)\sqrt{A}/\sqrt{2q{I}_{\rm{dark}}{GF}+4{k}_{B}T/{R}_{\rm{dark}}^{({\rm{PD}})}}$$
(7)

Even in studies presenting experimental detectivity assessments, reporting \({D}_{{theor}.}^{* }\) is encouraged to facilitate identifying deviations from ideality and allow cross-study comparisons. However, superlative claims regarding high \({D}_{{theor.}}^{* }\) compared with experimentally measured D* or \({D}_{\bar{B}}^{* }\) must be avoided, as they risk misrepresentation.

Speed of response

An important challenge in reporting and benchmarking the speed of response (Supplementary Note 3) of emerging photodetector technologies is that it generally depends on the applied voltage, load impedance, continuous-wave incident optical power and wavelength. Therefore, the speed of response should be characterized at several representative continuous-wave incident optical power levels within the LDR, with all the aforementioned parameters explicitly reported.

Characterization of speed of response in terms of the 10%–90% rise and fall times (τrise and τfall, respectively) on the application of a rectangular optical pulse (Fig. 5a, Supplementary Note 3 and Supplementary Fig. 8) requires the prior verification of the assumptions underlying these quantities. First, non-monotonic photocurrent transients near the pulse edges prevent the reliable extraction of τrise and τfall (for example, Fig. 5b,c and Supplementary Notes 19 and 20). Additionally, if a steady state is not reached after the edges of the optical pulse (for example, Fig. 5d), the extracted τrise and τfall are invalid in the context of general benchmarking. Therefore, for general benchmarking, photodetector studies should ensure that, after a pulse edge, the photocurrent waveform plateaus (for benchmarking purposes, we recommend considering that a steady state is achieved for variations of <1% from the peak value) over a timescale of ≥τr,p, where \({\tau }_{r,p}\stackrel{\scriptscriptstyle\mathrm{def}}{=}\left({\tau }_{\rm{rise}}+{\tau }_{\rm{fall}}\right)/2\). Moreover, it is crucial to verify that the photocurrent plateau during the optical pulse is consistent with the continuous-wave responsivity R0 for the applied optical power (Fig. 5a–c). However, in application scenarios involving short-pulse-train illumination, the photocurrent may not reach a steady state, yet the device may still exhibit high effective responsivity and practical utility. In such cases, transient response characterization under realistic excitation conditions and gain–bandwidth product evaluation may provide meaningful insights for application-specific benchmarking.

Fig. 5: Speed of response.
figure 5

a, Response to a rectangular optical pulse, alongside the illustration of rise and fall times. b, Response to a rectangular optical pulse with non-monotonic transients. c, Although shorter pulses may suggest that steady state is nearly reached, verification of the responsivity under illumination reveals a non-ideal transient. d, Example of incorrect extraction of rise and fall times for optical pulses that do not allow steady state to be reached. e, Response to an optical impulse. f, Bode plot of the responsivity of a photodetector behaving as an ideal linear first-order system under sinusoidally modulated light. ZL, load impedance; PD, photodetector.

If the speed of response is characterized by applying an ultrashort optical pulse (optical impulse; Fig. 5e, Supplementary Fig. 8 and Supplementary Notes 3 and 21) and quantifying the time τr,δ taken for the photocurrent to decay from 90% to 10% of its peak value following the impulse (Fig. 5e), it is crucial to verify that the ratio between the integrated photocurrent Qδ and the optical impulse energy Eδ (Fig. 5e) matches the continuous-wave responsivity. Reporting the energy and duration of the applied optical impulse is also essential.

If the speed of response is characterized in terms of the 3-dB frequency under variable-frequency sinusoidally modulated illumination (f3dB,sin; Supplementary Notes 3 and 15 and Supplementary Fig. 9)—or square-wave modulation with appropriate signal analysis (Supplementary Note 3)—it is crucial to report a clear plateau in responsivity as a function of frequency (with variations of ≤5%) spanning at least one frequency decade in the low-frequency region (Fig. 5f). Moreover, for benchmarking, the amplitude of the applied optical waveform must be specified alongside the continuous-wave optical power component.

Alternatively, if the 3-dB frequency is obtained by calculating the Fourier transform of the time-varying photocurrent in response to an optical impulse (defining f3dB,δ; Supplementary Note 3), it is essential to verify and report that doubling the optical impulse energy (for example, by increasing the impulse amplitude and maintaining its temporal profile) results in an output photocurrent that is identical in shape apart from proportional amplitude scaling, and that the extracted f3dB,δ remains unchanged.

If the time evolution of photocurrent with optical power can be described by a linear, first-order differential equation, then f3dB,δ and f3dB,sin are equal and can be calculated from time-domain measurements as ln[9]/(2πτr,δ) and ln[9]/(2πτr,p) (ref. 44). Moreover, under the same assumption, τr,δ and τr,p are equal. However, a common misconception in the emerging photodetector literature is that these relationships hold by default. In fact, although they are typically accurate for conventional photodetectors, discrepancies among f3dB,δ, f3dB,sin, ln[9]/(2πτr,δ) and ln[9]/(2πτr,p) may be as large as several orders of magnitude for emerging photodetector technologies15. This is because the different operating conditions used in measuring f3dB,δ, f3dB,sin, τr,δ and τr,p and possible effects (for example, charge transport and trapping) with diverse kinetics may invalidate the assumption that the time evolution of photocurrent under a time-varying optical signal can be described by a linear, first-order differential equation. Therefore, the aforementioned relationships among speed-of-response parameters cannot be assumed with emerging photodetector technologies until experimental validation is obtained on a case-by-case basis.

We note that measurements involving ultrashort, high-power light pulses are more prone to triggering nonlinear effects, whereas sinusoidally modulated illumination or rectangular pulses may prevent such effects for sufficiently small excitation amplitudes. Therefore, for general benchmarking, characterizing the speed of response via small-signal sinusoidally modulated illumination or rectangular pulses is recommended, with its reporting and interpretation bounded by the considerations discussed above. However, in specialized applications involving high-intensity ultrashort pulses, characterization should prioritize excitation with pulse width and peak power matching the application’s demands (despite possible nonlinearities) to meaningfully assess the response speed and enable relevant benchmarking, whereas small-signal measurements may still offer complementary insights.

Stability

Although the reliability of emerging photodetector technologies is critical to ensure their viability in real-world applications, reports on the stability of emerging photodetector technologies are scarce. By contrast, conventional optoelectronics adhere to standard temperature ratings—typically, 0 °C to 70 °C for standard-grade devices and –20 °C to 85 °C and –40 °C to 85 °C for extended-range and industrial-grade devices, respectively45. Their rated deployment times range from 2–3 years for low-end applications to up to ~20 years for higher-end applications45. Given the considerably less stringent characterization conditions typically adopted in studies covering emerging photodetector technologies so far, it is evident that aligning such technologies with real-world reliability objectives is a crucial aspect yet to be addressed.

Conducting stability tests over years or decades is impractical, making accelerated ageing tests essential (Fig. 6a,b). In the research stage of emerging photodetector technologies, a desirable target is to assess suitability for real-world use in a few weeks (~1,000 h) of accelerated ageing.

Fig. 6: Accelerated ageing tests.
figure 6

a, Test monitoring responsivity as a function of temperature, quantifying the device lifetime as the time when 80% of the initial responsivity is reached. b, Corresponding lifetime–temperature plot, from which the activation energy of the degradation mechanism is extracted. c, Overestimation of lifetime from an accelerated ageing test in the presence of higher activation energy at higher temperatures. d, Underestimation of lifetime from an accelerated ageing test in the presence of a change in activation energy only over a narrow, intermediate temperature range.

To assess reliability and readiness for commercialization, the community researching emerging photodetectors technologies is recommended to adopt standardized accelerated ageing tests already established for optoelectronics (that is, IEC-60749 (ref. 46 and other volumes in the IEC-60749 series), Telcordia GR-468-CORE47, JESD22 (ref. 48 and other volumes in the JESD22 series); AEC-Q100 (ref. 49) and ECSS standards50; Supplementary Note 22 and Supplementary Table 15). In approaching the application of these standards, researchers should ensure the predictiveness of their ageing tests by confirming that degradation mechanisms under real-world operating conditions align with those accelerated during testing. For instance, the functional dependence of the observed performance parameter versus time should remain consistent across normal operating and stress test conditions (Fig. 6c,d). Therefore, during the initial exploration of these standardized protocols for emerging photodetector technologies, it is imperative to also conduct tests at intermediate conditions to assess consistency with the degradation kinetics under the accelerated ageing conditions.

Although existing standards provide a solid foundation for the stability characterization of emerging photodetector technologies under temperature, humidity and bias stress, it is important to note that, unlike conventional inorganic semiconductors, emerging photoactive materials may also degrade under illumination. Therefore, evaluating the stability of emerging photodetectors under continuous or cycled illumination is also essential. Stability characterization under continuous illumination should use optical power levels within the photodetector’s LDR, and performance parameters should be measured at regular intervals to monitor degradation effects. For stability measurements under cycled illumination (with a modulation frequency below f3dB,sin), the photodetector current should be continuously monitored, and performance parameters should be characterized at both beginning and end of the stress experiment.

Finally, emerging photodetector technologies have the potential to address a wealth of applications that introduce novel scenarios (for example, seamless interfacing with the human body, biodegradability or edibility). Given the inherent limits of existing standards in this context, we encourage the research community to formulate new accelerated ageing protocols tailored to these innovative application scenarios.

Characterization, reporting and benchmarking in context

An important challenge arises from the diversity of photodetector performance parameters and the varying operational conditions used for characterization. Addressing this challenge necessitates presenting performance parameters under consistent operating conditions. This is essential to allow cross-study comparisons and the identification of genuine materials and device advances. For instance, given the dependence of all performance metrics on the applied voltage, it is essential for photodetector studies to report the device characterization at least at one common applied voltage (hereafter referred to as the prevailing voltage) for consistency and benchmarking.

To mitigate the data reliability challenge posed by potential device instability during characterization, it is strongly recommended to conduct and report a minimum of two repeated instances of current–voltage sweeps in the dark, spectral responsivities and frequency responses under sinusoidally modulated light as a sanity check for the same device when presenting such measurements. Additionally, measuring and reporting dark current and spectral responsivity data before and after noise measurements and time-domain characterizations of the speed of response are essential reality checks. Given the potential substantial device-to-device variability in early stage photodetector technologies, it is also crucial to avoid reliance on a single champion measurement. Instead, performance parameters should be reported with statistics derived from a representative batch of nominally identical devices. The exact batch size should be determined based on fabrication complexity and achievable yield. Although a batch of ten nominally identical devices is often a useful starting point, smaller batches may be necessary depending on fabrication scale-up challenges or low yield, in which case these limitations should be explicitly reported.

Another contextual challenge stems from the multitude of applications relevant to emerging photodetector technologies and the application-specific target ranges for performance metrics (Supplementary Table 16). Narrowly pursuing the maximization/minimization of a specific performance metric without considering this context may hinder adaptability for applications requiring that metric to be engineered in the opposite direction. For instance, in spectrally selective photodetector research, regarding ultranarrowband performance (FWHMR ≈ 10 nm) as a universal objective is inappropriate, as many applications require much larger responsivity passbands. Similarly, regarding ultrafast operation as a universal objective is inappropriate, as several photodetector applications do not necessitate high speed (for example, Supplementary Table 16). To address this challenge, studies presenting advancements in emerging photodetector technologies should identify their application context and the relevant required performance metric ranges.

Another nuance of the point above concerns avoiding emphasis on performance parameters that may not matter. For instance, many reports emphasize ultrahigh photoconductive gains, obtained, however, by devices with response times of up to minutes. In fact, response times on the order of minutes are unsuitable for most, if not all, applications (for example, Supplementary Table 16). Therefore, making superlative statements about metrics that may not matter should be avoided.

Finally, given the diverse expertise within the community researching emerging photodetector technologies, we acknowledge that certain studies may focus on innovations in materials and fabrication methods over the characterization of specific performance metrics. For instance, researchers developing highly innovative photodetector materials may lack the necessary equipment for comprehensive performance characterization. In these cases, we recommend adhering to the appropriate theoretical approximations discussed here, transparently presenting them as such, and benchmarking the resultant estimates solely against the literature relying on equivalent approximations.

Conclusions

This Consensus Statement tackles the need for accurate characterization, reporting and benchmarking of emerging photodetector technologies. By analysing definitions, misconceptions, context and challenges related to various performance metrics, we offer practical guidance and propose detailed guidelines to support consistent practices, accounting for the diversity of expertise and resources in the field. These guidelines will enhance understanding, evaluation and the real-world impact of emerging photodetector technologies.

Methods

This work builds on the outcomes of round-table discussions on photodetector characterization involving a team led by V.P. and involving T.D.A., B.N., D.N., M.C., R.H., G.G., J.L., E. Lidorikis, B.B. and J.J. These discussions took place in August–September 2022. Following these discussions, an outline for the Consensus Statement was drafted, spearheaded by V.P. with input from the team members. Subsequently, additional co-authors were invited based on criteria ensuring representation across diverse research streams in emerging photodetector technologies, inclusion of industry experts and geographical diversity. The manuscript was circulated among all contributing authors multiple times, with their inputs and feedback integrated iteratively until a consensus was reached.