Abstract
A method for high-accuracy analog optical computing based on interval arithmetic and the fixed-point theorem is considered. Two-variable simultaneous equations are studied to investigate the proposed method. An optical implementation is considered by the use of spatial coding of intervals, affine transformation, and image magnification. Computational simulation verifies the principle of the method.
© 1996 Optical Society of America
Full Article |
PDF Article
More Like This
Cited By
You do not have subscription access to this journal. Cited by links are available to subscribers only. You may subscribe either as an Optica member, or as an authorized user of your institution.
Contact your librarian or system administrator
or
Login to access Optica Member Subscription
Figures (6)
You do not have subscription access to this journal. Figure files are available to subscribers only. You may subscribe either as an Optica member, or as an authorized user of your institution.
Contact your librarian or system administrator
or
Login to access Optica Member Subscription
Tables (2)
You do not have subscription access to this journal. Article tables are available to subscribers only. You may subscribe either as an Optica member, or as an authorized user of your institution.
Contact your librarian or system administrator
or
Login to access Optica Member Subscription
Equations (16)
You do not have subscription access to this journal. Equations are available to subscribers only. You may subscribe either as an Optica member, or as an authorized user of your institution.
Contact your librarian or system administrator
or
Login to access Optica Member Subscription