Approximation capabilities of neural networks on unbounded domains.

2022 
There is limited study in the literature on the representability of neural networks on unbounded domains. For some application areas, results in this direction provide additional value in the design of learning systems. Motivated by an old option pricing problem, we are led to the study of this subject. For networks with a single hidden layer, we show that under suitable conditions they are capable of universal approximation in Lp(R×[0,1]n) but not in Lp(R2×[0,1]n). For deeper networks, we prove that the ReLU network with two hidden layers is a universal approximator in Lp(Rn).
    • Correction
    • Source
    • Cite
    • Save
    • Machine Reading By IdeaReader
    58
    References
    0
    Citations
    NaN
    KQI
    []