专利汇可以提供Compact integrated infrared scene projector专利检索,专利查询,专利分析的服务。并且A small and compact infrared scene projector allows for projection of infrared images from a variety of sources. The unit utilizes an external power supply to reduce the weight of the unit and has several input pathways such as an integrated RS-232 serial port and a video port for receiving RS-170 formatted video signals. The projector uses a digital signal processor to control its internal electronics and for on-board generation of pre-programmed infrared images. A processor electronics card and a scene generator electronics card communicate internally to generate images and to control a thermoelectric cooling device mounted to a semiconductor infrared emitter array to project flicker free, high resolution infrared images. Optics in the form of an interchangeable lens or a mirrored collimator allow for projection of a generated image onto a test object. Internal high speed memory and electrically erasable firmware, both externally programmable, allow for on-the-fly programming and self-sustaining and continued operation after disconnection from a separate programming computer. An external, user programmable interface allows for the download of commands and images into the scene projector. Once programmed by an external computer, such as a personal notebook computer, the scene projector can be placed in the field for live, self test of military electronics that rely on infrared sensors to make decisions.,下面是Compact integrated infrared scene projector专利的具体信息内容。
Having set forth the nature of the present invention, what is claimed is:1. An apparatus for displaying an infrared image, comprising:a. an infrared emitter array for converting electrical signals into an infrared image;b. an optical assembly for projecting said image upon a selected object;c. an electrical sub-assembly for generating timing and input-output control signals to said apparatus;d. a cooling element thermally connected to said emitter array for cooling said same;e. a processor assembly for generating image data utilized by said emitter array; and,f. wherein said emitter array, said optical assembly, said electrical sub-assembly, said cooling element, and said processor assembly are integrated within a single, self-contained module.2. An apparatus as recited in claim 1, wherein said processor assembly includes means for generating synthetic images.3. An apparatus as recited in claim 2, further including means for receiving externally created images for display by said apparatus.4. An apparatus as recited in claim 3, wherein said receiving means includes means for receiving an RS-170 formatted signal.5. An apparatus as recited in claim 3, wherein said means for generating synthetic images comprises a digital signal processor.6. An apparatus as recited in claim 5, wherein said electrical sub-assembly comprises firmware control logic for administering element intensity in said emitter array and control logic for supplying timing and control signals for receiving video information.7. An apparatus as recited in claim 1, further including means for receiving externally created images for display by said apparatus.8. An apparatus as recited in claim 7, wherein said cooling element comprises a TEC type of cooling device.9. An apparatus as recited in claim 7, wherein said electrical sub-assembly comprises firmware control logic for administering element intensity in said emitter array and control logic for supplying timing and control signals for receiving video information.10. An apparatus as recited in claim 9, wherein said cooling element comprises a TEC type of cooling device mounted directly to said infrared emitter array.11. An apparatus as recited in claim 1, wherein said cooling element comprises a TEC type of cooling device mounted to said infrared emitter array and wherein said apparatus further comprises a heat sink thermally connected to said emitter array.12. An apparatus as recited in claim 11, wherein said means for generating synthetic images comprises a digital signal processor.13. An apparatus as recited in claim 12, further including means for receiving externally created images for display by said apparatus.14. An apparatus as recited in claim 1, wherein said optical assembly comprises a reflective collimator for reflecting said infrared image onto a selected object.15. An apparatus as recited in claim 14, wherein said processor assembly includes means for mathematically generating images through algorithm processing.16. An apparatus as recited in claim 15, further including means for receiving externally created images for display by said apparatus.17. An apparatus for displaying an infrared image, comprising:a. means for emitting an infrared image;b. means for focusing said image upon a selected object;c. means for generating internal control and timing signals for said emitter means;d. means mounted to said emitting means for cooling said same;e. means for controlling said apparatus; and,f. wherein said apparatus is a single, self-contained module.18. An apparatus as recited in claim 17, wherein said control means includes means for synthetically generating images.19. An apparatus as recited in claim 18, further including means for receiving externally created images for display by said apparatus.20. An apparatus as recited in claim 19, wherein said receiving means comprises means for receiving an RS-170 formatted signal.21. An apparatus as recited in claim 19, wherein said control means comprises a microprocessor based assembly card.22. An apparatus as recited in claim 21, wherein said means for generating internal control signals comprises a firmware array.23. An apparatus as recited in claim 17, further including means for receiving externally created images for display by said apparatus.24. An apparatus as recited in claim 23, wherein said cooling means comprises a TEC type of cooling device.25. An apparatus as recited in claim 23, wherein said means for generating internal control signals comprises a firmware array.26. An apparatus as recited in claim 25, wherein said cooling means comprises a TEC type of cooling device and said cooling means is mounted directly to said emitting means.27. An apparatus as recited in claim 17, wherein said cooling means comprises a TEC type of cooling device and said cooling means is mounted directly to said emitting means.28. An apparatus as recited in claim 27, wherein said control means comprises a digital signal processor based assembly card.29. An apparatus as recited in claim 28, further including means for receiving externally created images for display by said apparatus.30. An apparatus as recited in claim 17, wherein said focusing means comprises a reflective collimator for reflecting said infrared image upon a selected subject.31. An apparatus as recited in claim 30, wherein said control means includes means for synthetically generating images through algorithm processing.32. An apparatus as recited in claim 31, further including means for receiving externally created images for display by said apparatus.33. An apparatus for displaying an infrared image, comprising:a. an infrared emitter array for converting electrical signals into an infrared image;b. an optical assembly for projecting said image upon a selected object;c. an electrical sub-assembly for generating electrical control signals received by said emitter array;d. a cooling element thermally connected to said emitter array for cooling said same; and,e. an embedded processor for generating image signals displayed by said emitter array.34. An apparatus as recited in claim 33, wherein said processor includes programming instructions for synthetically generating images through algorithm processing.35. An apparatus as recited in claim 34, further including means for receiving externally created images for display by said apparatus.36. An apparatus as recited in claim 34, wherein said processor comprises a digital signal processor.37. An apparatus as recited in claim 33, wherein said processor comprises a digital signal processor.38. An apparatus as recited in claim 37, wherein said cooling element comprises a TEC type of cooling device.39. An apparatus as recited in claim 33, wherein said optical assembly comprises a reflective collimator for reflecting said infrared image onto a selected object.40. An apparatus as recited in claim 39, wherein said processor includes programming instructions for synthetically generating images through algorithm processing.41. An apparatus as recited in claim 40, further including means for receiving externally created images for display by said apparatus.42. A method for generating synthetic infrared images, comprising the steps of:a. receiving a command function from an external computing device to synthesize a predefined infrared object;b. receiving a set of object descriptor variables associated with said predefined infrared object;c. mapping one or more recorded primary drawing functions to said predefined infrared object;d. synthesizing said predefined object; and,e. projecting said synthesized object.43. A method as recited in claim 42, wherein said step of synthesizing said predefined object comprises the steps of:a. executing said associated primary drawing functions;b. writing results of said drawing functions into a memory device;c. repeating steps a and b until said object is mathematically computed and saved; and,d. saving said mathematical results into a memory matrix modeling a pixel matrix of an infrared emitter array.44. A method as recited in claim 43, wherein said step of projecting said synthesized object comprises the step of writing said mathematical results in said saved pixel matrix onto said emitter array to project said synthesized object.45. A method as recited in claim 44, further including the step of interpreting said object descriptor variables to modify display of said predefined object in a predetermined manner to create an infrared scene.46. A method as recited in claim 45, wherein said steps of synthesizing said predefined object and repeated iteratively to create a continuous display of preprogrammed infrared images in said scene.47. A method as recited in claim 43, further including the step of interpreting said object descriptor variables to modify display of said predefined object in a predetermined manner to create an infrared scene.48. A method as recited in claim 47, wherein said steps of synthesizing said predefined object and repeated iteratively to create a continuous display of preprogrammed infrared images in said scene.
This application claims the benefit of filing priority under 35 U.S.C. §119 and 37 C.F.R. §1.78 of the co-pending U.S. Provisional Application Ser. No. 60/351,566 filed Jan. 24, 2002, for a COMPACT INTEGRATED INFRARED IMAGE SCENE PROJECTOR. All information disclosed in that prior pending provisional application is incorporated herein by reference.
FIELD OF THE INVENTION
The present invention relates generally to image projectors. In particular, the present invention relates to military type image scene projectors. In greater particularity, the present invention relates infrared image scene projectors.
BACKGROUND OF THE INVENTION
A good deal of military and communications weaponry today relies upon infrared imaging and acquisition techniques. For example, missile targeting systems rely upon templated infrared signatures to lock onto aircraft exhaust signatures, helicopter gun ships use tank heat signatures to lock onto targeted vehicles, and military aircraft navigation aids use infrared sensors to assist in auto-navigating through rough terrain and inclement weather. In more complex systems, such as satellite imaging, navigation element tracking, and forward looking infrared (FLIR) navigation systems, recorded infrared scenes are compared to live sensory data for real-time decision making. In general, infrared image sensing signatures of geography, buildings, and even people will likely become a major component in target acquisition and locking systems; and it is expected that various sorts of weaponry—missiles, artillery, navigation, and rifle targeting systems—will utilize infrared imaging in some form or fashion in the future.
Infrared image sensors detect electromagnetic radiation from a scene and output electrical or optical information extracted from the scene. Lenses associated with a selected imaging system focus infrared radiation upon one or more sensors so that acceptable signal capture is obtained. However, while image detection has been advanced due to improvements in infrared focal plane arrays (FPAs) and advanced optics design, a great deal of attention is now being drawn to the design of software for image recognition and decision making based upon received images. Sophisticated software analyzes received images and makes logical determinations based upon pre-established criteria, such as object templates, spectrum signatures (e.g. frequency signatures), and intensity to analysis.
Testing of infrared imaging systems can be as complex as the imaging design itself. Most production testing methods inject an electrical test signal for a sensor system to process and compare system component responses based upon an expected result. However, signal injection is not an optimum test method because the image detector with its related electronics and associated gimbals and optics are not tested. The optics and detector of the system are usually tested separately using a very simple static scene or test pattern, or with a very slow moving, but somewhat more complex series of synthetic scenes. This two-part test methodology does not perform an end-to-end test of the entire system and is inferior to providing live scene image testing data. Hence, as the quantity and complexity of such systems has increased the need for generating and projecting synthetic images for input into image sensors has also grown.
Several systems are now available capable of presenting realistic, detailed infrared images or “scenes” into the field of view of the sensors being tested for a selected system. For example, Kenyon U.S. Pat. No. 6,123,288, has developed a system to provide flicker-less projection of infrared scenes using a non-volatile electron beam addressed light-valve output. Leddy, U.S. Pat. No. 5,457,493, has developed a micro-mirror based image simulation system in which a digital micro-mirror reflects infrared energy from a separate infrared source onto a testing subject. And, Billingsley, U.S. Pat. No. 4,530,010, has designed a dynamic infrared scene projector using thin films of vanadium dioxide which are excited to produce infrared images.
However, each of these designs and others currently available in the industry suffer from several drawbacks, namely: (1) none of the currently available systems allow for easy mobile transport of a scene generator; (2) none allow for solid state generation of scene images on location and at reduced atmospheric pressures; and (3) none provide a self-contained, compact design that is capable of self-generation of images without external stimulus. Hence, military systems utilizing infrared sensory electronics cannot be easily tested in the field and do not include the ability for continuous live testing in airborne vehicles.
Moreover, one of the most desired testing methods for testing defense armaments and navigation systems is “built-in-self test (BIT). However, BIT is most useful if the testing hardware can be utilized in the field in the typical environment of the system. For example, if a scene projector could be integrated into a FLIR system, live pre-flight or in flight testing could be accomplished to ensure the integrity of the system's operation. BIT could also allow relatively easy refinements to existing imaging systems and allow the enhancement of on-board software recognition systems, thereby increasing the effectiveness of the current infrared sensory electronics.
Therefore, what is needed is a compact, self-contained infrared scene projector that can be configured to test infrared imaging systems in the field and during live operations.
SUMMARY OF THE INVENTION
It is the object of the present invention to provide a solid state, self-contained infrared scene projector.
It is another object of the present invention to provide an infrared scene projector that can internally create infrared scene images for projection onto a target testing object.
It is still another object of the current invention to provide an infrared scene projector that can receive image commands from an external computer for projection.
It is yet another object of the current invention to provide an infrared scene projector that can receive external video source imagery and project such source imagery.
It is still another object of the current invention to provide an infrared scene projector that can provide internal cooling for an infrared pixel array and thereby properly regulate infrared background temperatures of projected solid objects.
It is still another object of the current invention to provide an infrared scene projector that allows for interchangeable optics.
It is still another object of the current invention to provide an infrared scene projector that can be integrated with a mirrored collimator for multi-spectral projection of scene imagery.
In summary, the present invention is a relatively small and compact infrared scene projector. An external power supply powers the projector unit and an integrated RS-232 serial port on the projector allows for reception of commands from a connected personal computer. The projector utilizes a digital signal processor to control the internal electronics and for generation of preprogrammed infrared objects. A processor assembly card and a scene projector assembly card communicate internally to control a thermoelectric cooling (TEC) device and an electronic infrared projector array to project flicker free, high resolution infrared images onto a targeted sensor for testing. Optics in the form of interchangeable lens or a mirrored collimator allow for projection of an array generated image onto a live test object. Internal high speed memory and electrically erasable firmware, both externally programmable, allow for on-the-fly programming and self-contained operation. The projector can be co-located with military electronics that utilize infrared sensory imagery to provide field located built in self test (BIT).
Other features and objects and advantages of the present invention will become apparent from a reading of the following description as well as a study of the appended drawings.
BRIEF DESCRIPTION OF THE DRAWINGS
A scene projector incorporating the features of the invention is depicted in the attached drawings which form a portion of the disclosure and wherein:
FIG. 1
is an exploded perspective view of the bottom side (access panel side) of scene projector showing the primary physical components of the system;
FIG. 2
is a front perspective view of an assembled scene projector;
FIG. 3
a
is a perspective view of another embodiment of the assembled scene projector showing an attached tower and reflective fold mirror assembly in place of an optical lens;
FIG. 3
b
is a diagrammatic view of the scene projector including the tower and reflective fold mirror with an adjacent collimator;
FIG. 4
is a front elevational view of the assembled scene projector;
FIG. 5
is a side elevational view of the left side of
FIG. 6
;
FIG. 6
is a plan view of the top of the scene projector;
FIG. 7
is a side elevational view of the right side of
FIG. 6
;
FIG. 8
is a plan view of the bottom of the assembled scene projector;
FIG. 9
is an end view of the assembled scene projector;
FIG. 10
is a perspective cutaway view of the assembled scene projector showing the relative positions of the projector's housing, the DSP CCA, the Scene Projector CCA, a passive heat sink and TEC cooling device, the array CCA, and the optical subassembly;
FIG. 11
is a functional system diagram showing the primary functional subsystems of the scene projector and their inter-communication data paths;
FIG. 12
is a system component diagram showing functional communication paths between the primary electrical elements of the scene projector CCA;
FIG. 13
is a system component diagram showing functional communication paths between the primary electrical elements of the DSP CCA; and,
FIG. 14
is an optical element diagram of the optical subassembly used in the scene projector.
DESCRIPTION OF THE PREFERRED EMBODIMENTS
Referring to the drawings for a better understanding of the function and structure of the invention,
FIG. 1
shows an exploded view of the primary components of one embodiment of the infrared scene projector
10
. A primary housing or case
11
acts as a protective enclosure and support for the primary electronics assembly
12
which is mated to an infrared array assembly
13
. As shown in the Figure, screws affix the combined electronics assembly
12
and infrared array assembly
13
into housing
11
along an outer circumferential margin of the array assembly board
13
seated on an inner circumferential flange of the housing
11
. A lens mounting plate
14
is aligned onto dowels extending forward from the primary housing
11
with additional screws securing the plate
14
against the housing as shown, enclosing the electronics assembly
12
and the array assembly
13
within the housing
11
. Aperture
15
is located at the center of the lens mounting plate
14
to allow passage of the infrared emissions emanating from the seated infrared emitter array. The aperture
15
also serves as a receptor for lens assembly
16
which includes a threaded or smooth extension from its body for mounting onto the plate
14
. As shown in
FIG. 2
, the to assembled scene projector
10
occupies a rather compact and unobtrusive shape, having a typical thickness of approximately 3 inches and a total length measured from the foremost point on the lens assembly to the rearmost portion of the housing of only about 8 inches.
FIGS. 3
a-b
shows another embodiment of the scene projector
10
in which the enclosed electronics (
12
,
13
) and housing components (
11
) are outfitted with a tower and fold mirror in lieu of a refractive lens assembly
14
,
16
, and placed adjacent to a collimator assembly. In certain applications it is advantageous that the scene-projector
10
be positioned in front of a large aperture enclosed within a weather resistant housing. In such applications, a fold mirror and collimator assembly allows for testing structures to be positioned within an aircraft or vehicle for the purposes of mission readiness and target acquisition refinements. As shown in
FIG. 3
a
, an angled collimator mounting plate
17
supports a tower assembly
18
of multiple leg extensions
18
a-h
, and a fold mirror
19
is held by a suitable bracket
21
that allows for positioning of the mirror in an angle for reflecting a projected infrared image into an adjacent collimator. As shown in
FIG. 3
b
, a typical collimator assembly
22
might include, two curved mirrors, one convex and one concave, both suitably shaped to focus a projected infrared image onto a proximally located infrared sensor objective. As will be understood in the industry, various types of collimator lens and shapes can be manufactured to address various types of targeting sensor sizes and other physical configurations of infrared sensor electronics. It will also be noticed that the collimator mounting plate
17
is angled at approximately 12.32°. However offset angles of various ranges are anticipated to address various types of collimator and fold mirror configurations and sizes. The figure shows a Ritchey-Chretien type of collimator, although various configurations and types will be utilized. In the shown example configuration, both mirror surfaces are off-axis aspheric sections with an 8-inch clear aperture having a 36-inch focal length. An interface plate engages alignment pins to allow for positioning of the infrared array at the tilted image plane of the collimator.
FIGS. 4-9
show different elevational views of the refractive lens embodiment of the infrared projector
10
.
FIG. 8
additionally shows an access plate
26
that allows for access to the internal wiring harness (not shown) connecting the internal electronics assembly boards to the projector's external port connectors.
FIG. 9
shows external connectors, namely, a standard RS-232 9 pin serial port connector
28
, a 10 pin recessed pin grid connector
29
which provides power and grounding for the projector, and a video input jack
30
that provides a port for receiving RS-170 video signals.
FIG. 10
shows a perspective cut away view of the assembled scene projector showing relative positions of various critical components for the scene projector
10
. The projector housing
11
holds the passive heat sink
32
via screws protruding through each heat sink cooling fin, allowing for isolation of the heat conducted away from the array emitter
51
from the primary electronics assembly
12
. Bracket
31
holding electronic assemblies
34
and
36
(not visible) is secured into array emitter CCA
13
with
4
screws extending through the emitter CCA
13
and into the bracket
31
. The assemblies
34
and
36
are secured to the bracket with flat pan screws. A TEC (Thermoelectric Cooler) device
33
is positioned between heat sink
32
and infrared array assembly board
13
and affixed to a flat portion of the heat sink
32
with thermally conducting pads and to the bottom of the actual array emitter. As can be seen in the figure, connectors are suitably positioned to engage each of the boards electronic internal connectors and provide for electrical communications between the boards over prescribed lines. A rectangular aperture in the printed circuit board holding the infrared array emitter
51
allows the TEC cooling device to be mounted directly onto the ventral surface of the ceramic semiconductor package of the emitter array
51
, which is affixed thereto with suitable heat conducting pads. The heat sink
32
is also adhesively affixed to the heat transference side of the TEC to promote transfer heat emissions generated by the infrared emitter away from the emitter and against the inner surface of the housing
11
Heat is therefore conducted away from the emitter
51
and the primary electronic assembly
12
through the housing
11
top and bottom surfaces. As those skilled in the art will appreciate, establishing proper controls for background infrared emissions is an important component in proper infrared scene generation. The disclosed arrangement of the TEC cooler and suitably sized heat sink allows for sufficient dissipation of heat energy from the infrared array emitter such that, under proper electrical control as will be discussed, background energy emissions are controlled.
Thermoelectric cooling devices such as the TEC cooler disclosed herein
33
are primarily electric heat pumps used for the removal of heat from one side of a TEC device to another side. Each side is sometimes referred to as the hot and cold sides of the TEC device. In the disclosed embodiment, the cold or cooling side of the TEC device is affixed to the rear or underside portion of the infrared emitter array and the hot or heat dissipation side is located on a side opposite from the cooling side and against a forward-most, flat portion of the passive heat sink 32. A suitable TEC cooler utilized in the disclosed design is a FRIGICHIP FC series type available from Melcor, Inc. The passive heat sink
32
is made from suitably conductive metal alloys such as conductive aluminum which satisfactorily dissipates heat transferred by the TEC cooler. As infrared emitter technology advances, it is anticipated that a passive heat sink design may, in and of itself, provide sufficient cooling, thereby obviating the need for thermoelectric cooling.
The infrared array emitter
51
can be comprised of any industry standard matrix addressable infrared emitter (albeit with some electronics reconfiguration for each model). The current preferred embodiment utilizes analog generated input signals to address image pixel intensity requirements, although the inventors contemplate that a purely digital based infrared emitter will be available in the future. A suitable emitter for the current design is offered by Honeywell, Inc. under the brand name Brite 128-512 Infrared Emitters. These types of emitters directly radiate energy from underlying pixel addressable points on a silicon wafer. A matrix of 128×128 pixels form a matrix of gray body emitters with an infrared radiance that varies as a function of an applied voltage. These types of Honeywell emitters generate infrared radiance at individual pixel locations by passing current through a thin film resistor suspended above a polarized substrate. The current through the emitter resistor is a function of the gate voltage stored in the hold capacitor connected to the gate allowing for very low power operation by which pixel emissions vary within a wide dynamic range. Moreover, these types of infrared emitters have high vacuum integrity because the pixels are fabricated using low vapor pressure materials, such as nitride and oxides, with bulk properties having melting temperatures in the 1500-2000 k temperature range. Hence, each pixel emits a high contrast infrared point with very high thermal stability that performs over a range of atmospheric pressures, such as may be experienced in airborne applications.
As partially shown in
FIG. 10
, connectors on the infrared emitter array assembly board
13
provide for electrical signal connections between various electronic assembly boards, and wiring harness
38
provides electrical signal connectivity from external connectors
28
-
30
to the various boards.
Referring now to
FIG. 11
, the system consists of four functional components: (1) a DSP circuit card assembly (“CCA”)
36
; (2) a scene generator circuit card assembly
34
; (3) an infrared emitter array CCA
13
; and, (4) an optical assembly
53
. The scene generator CCA
34
provides power
41
to other functional elements in the system. A RS-232 port
43
provides serial information input into the DSP CCA
36
to receive computer commands from a graphical user interface running on a separate personal computer (not shown) and also provides for reprogramming of internal firmware and downloading of DSP operational commands. A connector
30
(see
FIG. 9
) allows for RS-170 signals
42
to be received by the scene generator CCA
34
, which are then separated into digital video data
64
and video synchronization information
67
. A digital signal processor or “DSP” integrated semiconductor
46
provides video data
47
for conversion through a digital analog converter
48
into analog signals
49
received by infrared emitter array
51
for infrared image generation
52
. The DSP
46
provides direct video feed from the RS-170 signal into the digital analog converter
48
or, alternatively, the DSP can generate its own infrared images through algorithm computations stored in FLASH memory and in association with logic stored in firmware FPGA
56
(Field Programmable Gate Array). Also, an RS-232 source
43
can provide a bit mapped scene image to be held in high speed memory on the DSP CCA
36
such that it can then be transferred directly to the infrared image emitter array
51
. A TEC cooler
58
provides cooling to the array emitter
51
as already discussed and thermistor
99
outputs a feedback signal to analog digital converter
54
that is read by DSP integrated circuit
46
. In response, the DSP outputs commands through the FPGA interface
56
to control the TEC cooling rate through TEC control signals
57
.
FIG. 12
shows more specific details regarding the function and operation of individual components located on the scene generator CCA
34
. A buffer
71
normalizes voltages compatible with the CCA's voltage levels from the inputted RS-170 signal
42
, and a video sink separator
72
separates the video synchronization signal from the RS-170 buffered signal to establish a timing indicator signal for use by pixel clock generator
73
. The pixel clock generator includes logic to send appropriate clock signals via a pixel clock input line to a 12-bit analog to digital converter
74
. “Backporch” signal
76
from separator
72
is used to offset-normalize the video information
78
via DC restorer
77
to allow for proper digital conversion. Video information
78
is then converted from analog to digital via 12-bit video analog to digital converter
74
and saved in a 256K×18 FIFO
79
memory for interim storage of individual interlaced signal fields. The pixel clock generator
73
generates a field storage signal at approximately a 60 Hz rate so that interlaced fields can be combined into a non-interlaced frame by the DSP CCA
36
. The pixel clock generator
73
utilizes external phase loop logic
81
in conjunction with the pixel clock generator logic
73
contained within the FPGA
56
to produce the appropriate clock signals. The combination of these video and logic elements
82
allows for the continual conversion of RS-170 source signals, either in color or black and white, to be continually digitized and processed by the DSP CCA
36
. As will be explained in more detail, the DSP CCA
36
has available one field of video stored in the FIFO
79
for processing in accordance with prior downloaded instructions from a separate user software package or prior loaded internal processing instructions. Due to the speed of the DSP processor, processing of individual video fields stored in the FIFO
79
occurs well prior to the availability of the next interlaced video field presented by the external RS-170 signal source, thereby permitting the combining of interlaced fields into a non-interlaced image and the timely transference of the image to the emitter array
51
.
FIFO interface
83
provides a signal to the DSP CCA
36
signaling whether or not the FIFO
79
contains a complete field and is available for transference to the DSP CCA. FIFO interface
83
also sends control signals to the FIFO
79
instructing the FIFO to load digital video from the 12-bit video analog-to-digital converter
74
. Core interface logic
84
provides interface logic to accept control signals from the DSP CCA
36
via multiplexed buses
86
-
88
. Bus lines
86
-
88
are physically identical on the scene generator CCA
34
and on the DSP CCA
36
. However, the digital signal processor
46
utilizes chip selects to multiplex values on the bus
86
-
88
to present data and receive data from appropriate integrated circuits such as, for example, the FIFO video frame information
79
and transference of core logic instructions to core interface
84
on FPGA
56
. Bus
86
-
88
is utilized to transfer 32 bit data information to data buffer
91
and matrix addressing array buffer
92
. FPGA
56
also includes digital to analog conversion interface logic
95
that provides control signals to the 32 bit digital analog converter
93
. Those skilled in the art will understand, therefore, that the 20 bit address bus, the 32 bit data bus, and the control bus
86
-
88
are a multiplex set of lines circumscribing paths on both the scene generator CCA
34
and the DSP CCA
36
. By using a chip select function, the DSP CCA
36
can receive interlaced video field data from FIFO
79
and combine it with a second interlaced field to produce a non-interlaced display image frame for writing to the 32 16-bit DAC
93
in preparation for display.
A 32-bit data buffer
91
is used for signal fan out to the 32 16-bit DACs
93
. Pixel intensity information is retained in the 32 16-bit DAC
93
and presented to the infrared emitter array
51
via eight analog input lines
77
. These analog input values are held and the DSP CCA
36
energizes the appropriate address lines. A TTL-to-CMOS buffer
92
is used to convert the DSP CCA
36
3.3V TTL logic levels to 5V CMOS logic required by the infrared emitter array
51
. Additional address strobes are replicated until all of the pixels in the array
51
have been loaded with analog intensity information. Although a 128×128 pixel configuration is shown, those skilled in the art will understand that a scalable addressing scheme may be utilized to address larger arrays such as a 256×256 or 512×512 pixel arrays.
The TEC cooler
58
is mounted to the underside of the array
51
to provide cooling as previously discussed. A TEC cooler drive
96
receives constant frequency signal from a pulse width modulator
97
which varies its duty cycle in accordance with modulation logic in the FPGA
56
to control the rate of cooling by the TEC cooler
58
, and which in turn is controlled by command signals provided by the DSP CCA
36
through the core interface logic
84
. A thermistor measurement circuit
98
receives sensory information from a temperature sensitive thermistor
99
which is then converted into digital information through analog to digital converter
54
on the DSP CCA
36
. Digital signal processor
46
then interprets digital values of the thermistor measurement circuit
98
and sends appropriate control signals to the FPGA
56
to control the modulator
97
. Hence, through this feedback communication strategy, the TEC cooler
58
can be controlled with a high level of granularity that allows for control of the background infrared emissions of the infrared emitter array
51
, thereby enhancing infrared image generation integrity.
Module interface
101
comprises an external connector
29
and wiring harness (not shown) providing appropriate +5V, +12V, and −12V input voltages and returns. The module interface
101
also comprises the RS-232 pin connector
28
that allows for direct connection into the DSP CCA
36
. In order to minimize feedback noise potential, a separate pixel power line
102
provides +5V direct input and return into the infrared emitter array
51
. If common +5V lines were utilized to power the infrared emitter array
51
, a possibility of switching noise generated by the DSP CCA
36
could be propagated through power lines into the array CCA
13
and provide undesirable interference with the infrared emitter's operation.
Referring now to
FIG. 13
for a better understanding of the operation of the DSP CCA
36
, a TMS320C6211 digital signal processor
46
provides the primary processing functions of the DSP CCA
36
. Texas Instruments is the manufacturer of the TMS320 series of digital processors two of which, the TMS320C62211 (non-floating point) and the TMS320C6711 (floating point) digital signal processors, may be utilized in the instant described design. However, those skilled in the art will understand that any type of processor having the capability of executing instructions and algorithmic processes at sufficient speeds may be utilized. For example, the inventors anticipate that a general purpose microprocessor such as Intel's 386 16-bit line of microprocessors could be utilized in place of herein described digital signal processor. The DSP CCA 36 is a fairly self-contained CCA and may be utilized in various other systems. The actual DSP operates at 150 MHz and uses an internal 100 MHz system bus, and the speed of the DSP CCA
36
is approximately 1200 MIPS or 600 MFLOPS. The DSP CCA includes 256K×16 bit FLASH memory
111
and 64K×32 bit asynchronous static memory (ASRAM)
112
for DSP program store and image calculations. The CCA
36
also utilizes several high speed and low speed (100 MSPS, 100 KSPS) analog to digital converters (not shown) controlled by a XLINX FPGA through control lines
114
.
The procedures by which the TMS32ODSP, or an alternative microprocessor, are programmed will not be discussed in as much as programming kits and code compilers are readily available and well understood in the industry. For example, Texas Instruments offers a software developer's kit called a “Code Composer Studio Development Environment” for the TMS320 line of DSPs and Intel sells similar developer kits for its line of microprocessors allowing for DSP code assembly and compilation.
The DSP
46
communicates to the scene generator CCA
34
via external memory interface EMIF
121
and via connectors
116
and
117
. Other communication links such as the RS-232 link
43
, a controlled area network link (not shown), and a host port interface (HPI)
130
can also be invoked for other types of communication and controls. Currently the EMIF is a little endian format, but other suitable formats may be utilized. Flash memory
111
includes FPGA reprogramming code
119
to allow for reprogramming of the XLINX FPGA
113
and includes a protective portion of flash memory for boot up code
125
. The DSP
46
and the XLINX FPGA
113
communicate with the scene generator CCA
34
through the external memory interface
121
. The DSP
46
utilizes four external address spaces in its external memory interface
121
, namely; CE
0
, CE
1
, CE
2
, and CE
3
. CE
0
is used to access static RAM
112
and CE
1
is used to access FLASH memory
111
, registers on a separate PLD (not shown), and registers on the XLINX FPGA
113
. The XLINX FPGA
113
decodes each of these address spaces into four sub address spaces to allow a total of eight addressable spaces. Therefore, the XLINX FPGA may utilize one of its sub-address enables using the top two address bits (ADR
21
and ADR
20
) from which chip enables CE
2
or CE
3
may be activated by the DSP
46
via the EMIF
121
. The scene projector CCA
34
generally utilizes 5 of the 8 sub addresses to decode logic signals.
The XLINX FPGA
113
includes all the necessary logic to decode and control various peripherals on the DSP CCA
36
. This includes the Analog to Digital Converter (ADC)
54
for receiving thermistor measurement signals
98
, DSP external interrupt control registers, such as interrupt pole, mask, and status registers, and an optional fast ADC, other optional fast and slow Analog to Digital Converters (DAC), LED control registers, clock control registers, and FPGA configuration status registers.
Communications between the DSP CCA
36
and the scene generator CCA
34
occur through CE decodes over the address bus, the data bus of the EMIF, and the read/write control signals of the EMIF bus. The CE lines (see Figure) decode whether the EMIF bus lines are being utilized for communication between the DSP
46
and the scene generator FPGA
56
, the DSP and the scene generator CCA digital analog converters, or the FIFO
79
. Transceivers
122
and
123
insure error free communications across connectors
116
and
117
between the DSP CCA
36
and the scene generator CCA
34
. Also, a separate programmable logic (PLD) device on the DSP CCA (not shown) controls the FLASH memory
111
, sets the timing input for and enables reprogramming of the XLINX FPGA
113
, contains XLINX FPGA
113
read back signals, and generates a 10 MHz clock for the use of an optional controlled area network controller.
FIG. 14
shows an optical element configuration suitable for the herein described scene projector. Various types of optical assemblies and optical elements may be utilized to project an image emitted by the infrared emitter
51
onto various objectives. The shown optical assembly is optimized to project an image into a selected FLIR objective lens. It will be understood that varying optical assemblies could utilize threaded or bayonet type mounts to facilitate the interchangeability of lens on mounting plate
14
to allow for rapid reconfiguration of the infrared projector
10
to suit different situations.
Inasmuch as optical assemblies and techniques for combining various optical elements to produce suitable focal lengths and fields of view are well known and understood in the industry, a detailed description of individual optical elements will not be provided. Nonetheless, an example configuration is shown in
FIG. 14
which has been used by the inventors and is suitable for the herein described types of applications. As shown, individual pixel elements of the infrared emitter array
51
emanate from plane position
131
and are refracted by zinc selenide optical element
132
. Element
132
has shaped properties pursuant to the values shown and is suitable for infrared refraction. Infrared image rays strike a second germanium optical element
133
and are again refracted to project an infrared image into a test article objective. Element
133
has optical properties as shown in the Figure. A typical distance of the optical element
132
and infrared source
131
is 16.4562 mm and the distance between Element
132
and Element
133
is 79.0 mm.
In operation, the scene projector
10
can produce infrared images through three primary methods: (1) it can project a received RS-170 video signal which is refreshed at television video rates; (2) it can project a preloaded bit-map image received via the RS-232 communications link; or (3) it can construct synthetic images from parameters received from a separate user interface running on a serially connected personal computer or other computing device. Preloaded images, either through parameter description or as a bit-mapped image, can also be saved in memory (
111
or
112
) and projected on demand. In fact, a demonstration program using stored images or a preloaded test program can be created by storing such images or parameters and accessing them through an iterative display loop program.
Creation of synthetic objects are accomplished by transferring inputs from an operator via a host interface to the scene projector
10
, and executing algorithms stored in memory using transferred variable values to create images. A separate application program running on a personal computer allows a user to send command functions to the scene projector instructing it to, for example, be ready to receive a bit mapped image for display, prepare for firmware programming, start displaying RS-170 images, begin receiving object descriptor information to calculate and display synthetic image objects, etc. Selected functions are communicated to the DSP in a binary string of digits as a packet in a form as shown in Table 1:
TABLE 1
Command Function and Argument Communication Format
Names of Binary
Fields of Each
Packet in sequence
Description
Byte Count
First 16 bits indicating the length in bits
of the packet.
Command Op. Code
Next 8 bits defining the current function.
Data
Argument for Command op. code function
having a variable length depending upon selected
op. code function.
CRC
Fixed length of 32 bits allowing for cyclical
redundancy checking.
Sequence
Unique 8 bit number identifying each transferred
packet of information to allow for matching of
acknowledgement responses from DSP.
In the situation in which a command function calling for an object to be synthesized and displayed by the scene projector is transferred to the DSP, the data argument would consist of image creation parameters called “object descriptors.” The PC application program includes a graphical user interface or “GUI” that allows for easy input of the object descriptors via predefined fields. These defined object descriptors are then grouped into the argument of the command function packet as part of the Data Field and transferred to the DSP CCA
36
via the serial RS-232 connection. The format and parameters of the object descriptor data portion is as shown in Table 2 below. The inventors envision a multitude of types and quantities of description elements, but
13
pre-defined parameters seem to satisfactorily describe most synthetic images. Since the type and operation of a graphical user interface is not essential for a complete understanding of the operation of the scene projector
10
, and since combining a list of input object parameters into a binary data packet and transferring it to the memory of a computing device is well understood in the industry and may be accomplished in a variety of ways, further description of such interfaces will not be described. The essential capability of the remote user interface is that it be able to transfer binary files and bit mapped images via the RS-232 link, and that the creation of the object descriptors be created and transferred into memory elements
111
or
112
. Object descriptors shown in Table 2 below are satisfactory for most synthetic images:
TABLE 2
A Suitable Object Descriptor Packet
Descriptor Variable
Name
Description
1. X Offset
No. of array pixels in X (±) direction by which
the object is offset from center of the emitter
array.
2. Y Offset
No. of array pixels in Y (±) direction by which
the object is offset from center of the emitter
array.
3. X Sweep Rate
Rate at which the object is moved from right to
left or left to right. Rate is based upon an internal
clock frequency.
4. Y Sweep Rate
Rate at which the object is moved from top to
bottom or bottom to top. Rate is based upon an
internal clock frequency.
5. X Sweep Range
Limit or bounds of a sweeping range in X
direction measured in no. of array pixels.
6. Y Sweep Range
Limit or bounds of a sweeping range in Y
direction measured in no. of array pixels.
7. Rotation Angular
Speed at which object rotates about center of
Rate
array from vertical (as altered by X and Y
Offsets) in radians/sec.
8. Rotation Angle
Set angle in radians by which an object is to be
rotated.
9. Object Type
Selection of a predefined and preprogrammed
object shape.
10. Object Size
Size of object in no. of array pixels.
11. Orbit Angular
Rotation rate about center of array (as altered by
Rate
X and Y Offsets) in radians/sec.
12. Orbit Radius
Radius dimension from center of array (as altered
by X and Y Offsets) in array pixels.
13. Temp. Difference
Temperature of emitted object in Degrees Celsius
above ambient temperature.
Upon transmission of a command function packet calling for the creation of a defined image, the DSP
46
uses the received descriptors to set the initial conditions of the image and to set a rate timer that regulates the motion of the object, if any. The DSP then calculates the location of the object origin in terms of two dimensional Cartesian coordinates. After calculating the sweep position and orbit angle from the sweep rate and orbit angular rates respectively, the origin is computed utilizing the specified sweep position, sweep range, orbit angle, and orbit radius. Once the origin of the object is determined, the object is drawn in static memory
112
as a matrix representing individual pixel elements of the array
51
.
Various predefined objects are preprogrammed into the DSP's memory
112
for execution upon receiving a recognized object type in the descriptor data (see parameter
9
). The selected object type at the required size and specified delta temperature is then drawn using a combination of three basic drawing functions (not to be confused with operation command functions). The primary drawing functions and their associated arguments are a “plotline” (x
0
,y
0
,x
1
,y
1
, width), a “plot circle” (radius), and a “fill” (color). Each object is drawn using a combination of these primary drawing functions calculated by the DSP in association with other user supplied inputs from the scene projector's GUI. Each drawing function has an initial configuration and is scaled using the descriptor parameters to form the selected object. For example, a bar object might be invoked, a 4-bar object, a circle object, a triangle object, or an alignment object. Solid objects such as the triangle and circle are filled with a color that is based on a user inputted temperature differential. If rotation of an object is required then the rotation angle and rotation angular rate are provided in the object descriptor data, which are used to compute the next required image in a movement sequence. This is done using a polar coordinate system in which the coordinates of key features of the object are first computed and then the object is drawn in the same manner as in the original object position. After an object image is drawn by the DSP and saved in memory, the object image information is written to the infrared array. In the event that the drawn image moves beyond the 128×128 array pixel matrix bounds, the portion of the object still within the boundary of the pixel matrix is displayed. This display strategy allows for the replacement or upgrading of the 128×128 array to larger arrays such as 256 or 512 pixels without rewriting the DSP instructions. Below, a description of how each primary object is drawn by the DSP is given in terms of the primary drawing functions. Obviously, as new objects are added over time, new functions may also be created to efficiently draw each object.
The bar object is essentially a 7:1 aspect ratio bar constructed from the line and fill functions. Based on the user specified size described in terms of number of pixel elements, two horizontal lines are drawn and two connecting vertical lines having a length 7 times that of the bar width are inserted between their endpoints. The resulting rectangle is then filled with the fill level or “color” based on the received delta temperature parameter.
The 4-bar object is created using a four 7:1 aspect ratio bar equally spaced by an equidistant amount. The perimeter of the resulting pattern is a square which is then centered in the array and each bar is then constructed as with the single bar method, but with lines added and a fill color drawn.
The circle object uses the size specified by the operator in parameter No. 10 as the radius of a circle where the units of the radius is in number of pixel emitters of the array. The circle function is used to draw the circle object and then the fill function is applied.
A triangle is created using three line functions and applying the fill function. The size of the triangle is specified by the user and determines the distance in emitters from the center of the array to the vertices of an isosceles triangle.
Lastly, the alignment object is generated by combining four concentric circle functions and two orthogonal lines through the center point of the circles. The temperature level of the lines are based on the temperature delta selected by the operator.
While the above four objects have been developed using the described three primary drawing line functions, it will be understood that additional drawing functions and objects will likely be developed depending upon the evolution of testing applications and additional testing refinements.
While we have shown our invention in one form, it will be obvious to those skilled in the art that it is not so limited but is susceptible of various changes and modifications without departing from the spirit thereof.
标题 | 发布/更新时间 | 阅读量 |
---|---|---|
基于分视控制的斜视与视功能训练阅读器及其调设方法 | 2020-05-11 | 155 |
汽车爆胎安全稳定控制方法 | 2020-05-12 | 994 |
图像互换格式图片的生成方法和移动终端 | 2020-05-14 | 682 |
图片显示方法、装置、计算机设备及存储介质 | 2020-05-16 | 394 |
全息展示窗的交互方法、装置以及电子设备 | 2020-05-12 | 516 |
一种多路视频预处理方法 | 2020-05-20 | 546 |
一种网页的处理方法及相关设备 | 2020-05-20 | 329 |
高精度自动光学检测方法和系统 | 2020-05-14 | 32 |
自动光学检测方法和系统 | 2020-05-21 | 941 |
实现量子计算机、量子通讯、裸视4D全息电视机系统的方法 | 2020-05-21 | 570 |
高效检索全球专利专利汇是专利免费检索,专利查询,专利分析-国家发明专利查询检索分析平台,是提供专利分析,专利查询,专利检索等数据服务功能的知识产权数据服务商。
我们的产品包含105个国家的1.26亿组数据,免费查、免费专利分析。
专利汇分析报告产品可以对行业情报数据进行梳理分析,涉及维度包括行业专利基本状况分析、地域分析、技术分析、发明人分析、申请人分析、专利权人分析、失效分析、核心专利分析、法律分析、研发重点分析、企业专利处境分析、技术处境分析、专利寿命分析、企业定位分析、引证分析等超过60个分析角度,系统通过AI智能系统对图表进行解读,只需1分钟,一键生成行业专利分析报告。