Norton's theorem (along with its dual, Thevenin's theorem) is a fundamental theorem in electrical engineering used for circuit analysis. It states that any linear electrical network containing independent and/or dependent voltage and current sources and resistors can be replaced by an equivalent circuit consisting of a single current source and a single parallel resistor connected across the load. The key condition for Norton's theorem (and Thevenin's theorem, superposition, etc.) to apply is that the network must be
linear. A linear network is one where the relationship between voltage and current in all its components (resistors, inductors, capacitors, and sources) is linear. This means:
- Homogeneity: If the input (voltage or current) is scaled by a factor, the output is scaled by the same factor.
- Additivity (Superposition): The response to multiple inputs is the sum of the responses to each input individually.
Components like ideal resistors, inductors, and capacitors are linear. Diodes, transistors, and many other semiconductor devices are examples of non-linear components. Therefore, Norton's theorem is true only for linear networks.