a project report · a project report by abdul rahman s. 720315114003 ajay hariharamanian k....
TRANSCRIPT
ACOUSTIC FENCING
A PROJECT REPORT
by
ABDUL RAHMAN S. 720315114003
AJAY HARIHARAMANIAN K. 720315114004
DINESH SHANKAR P. 720315114030
MANJU V.M. 720315114063
GUIDED BY
Prof.R.MOHAN RAJ M.E.,,
Assistant Professor,
Department of Mechanical Engineering,
Akshaya College of Engineering and Technology
iv
ABSTRACT
Fence is physical barrier placed as symbol of restriction for living beings
into particular areas; in agriculture it is a visible note that fencing is done in order
to prevent the entry of animals into the fields. Acoustic fence is a sound based
non-physical barrier for animals, which stop them from entering into the crop
fields. By means of replacement of electric fence with acoustic fence, the death
rate of animals can be reduced. Sound is utilized here to make the animals stay
away from crop fields.
TABLE OF CONTENTS
CHAPTER
NO. TITLE
PAGE
NO.
ABSTRACT iv
1. INTRODUCTION 1
2. LITERATURE REVIEW 3
3. INTRODUCTION TO SOUND 6
4. INFRASOUND 7
4.1 RESONANCE 8
4.2 STUDY ON ELEPHANTS AND RESONANCE 8
4.2.1. How often will an elephant use infrasonic calls? 9
4.2.2. Travelling of infrasonic elephant calls. 9
4.2.3. Analyzing infrasonic calls 10
4.2.4. A unique feature of infrasonic elephant calls which 10
distinguishes them from other infrasound that may
be recorded.
5. METHODOLOGY. 12
5.1 PRODUCTION OF INFRASOUND USING
PUMPING ACTION 12
5.2 INFRASOUND GENERATION USING ARDUINO 13
5.3 INFRASOUND GENERATION USING CLOCK PULSE 14
5.4 PRODUCTION OF INFRASOUND USING LAB VIEW 16
6. PROJECT BRIEF 17
6.1 IMAGE PROCESSING 18
6.1.1 Template creation 18
6.1.2 Image from webcam 18
6.2 DESCRIPTION OF FRONT PANEL 18
6.3 BLOCK DIAGRAM OF IMAGE PROCESSING 20
6.3.1 IMAQ 21
6.3.2 Vision acquisition 21
6.3.3 Vision assistant 23
6.3.4 Block diagram inputs 25
6.3.5 Block diagram outputs 26
7 CONCLUSION 28
8 COST ESTIMATION 29
1
CHAPTER 1
INTRODUCTION
A living creature needs food to survive. After all, we human beings live,
struggle and earn so much to fill our stomachs. Agriculture, which pours healthy
lives to people, is in problem, as we are all familiar with headlines like ―crop
fields destroyed by wild animals!‖ and so on. A fence is a structure that encloses
an area, typically outdoors, and is usually constructed from posts that are
connected by boards, wire, rails or netting. A fence differs from a wall in not
having a solid foundation along its whole length. Alternatives to fencing include
a ditch (sometimes filled with water, forming a moat).Though there are many
kinds of fences practiced, still we are not able to control the intruding of animals
into the crop fields. Bringing a solution to this is our project. The application of
infrasound in agricultural field can bring a vast change. There is a two way
advantage, as the crop field is made safe by restricting the entry of animals,
similarly the animals are safe, they need not be harmed, thereby giving equal
importance to lives of animals, humans and crops too.
Several surveys conducted on natural disasters shows that death rate of
animals are much lower than that of the humans . Natural disasters like
earthquake, hurricane, tornado, volcanic eruption leads to the production of
vibration at infrasonic level. It’s been myth for years that how animals could
predict and escape from the disasters, now science behind it is revealed that
animals can hear and sense the infrasound. The capability of hearing and sensing
the infrasound makes the animals to vacate from the place of disaster occurring.
Still several methods are used to shoo away the animals from their restricted
places. But while following these methods, there are harms caused to animals as
2
well as to humans in some way or the other. To rectify this problem we believe
infrasound can be a solution. By generating infrasound, we make animals to get
frightened about stepping towards the place from where the sound is originating.
By using Acoustic fence we can avoid the disadvantages of paling,
chemical and electrical fences, and saving the lives of humans and animals.
3
CHAPTER 2
LITERATURE REVIEW
(I) Gary W. Hicks Larry R. McDonald. et.al., (2003) [1], methods and
apparatus determine if an underwater intruder passes under a protective boundary. A
sonar sensor system comprises a plurality of sonar sensor modules that are spaced on
a protective boundary. A sonar sensor module comprises a sonar transducer (sonar
array) that is characterized by an omni-directional radiation pattern that may overlap
an omni-directional radiation pattern of an adjacent sonar sensor module transducer.
The sonar sensor module collects sonar data such as range information of the target in
relation to time. A central processor obtains the sonar data from each sonar module
through a telemetry link. The central processor processes the sonar data from the
plurality of sonar sensor modules in order to determine an estimated path of the target
and may determine if the target should be considered as a threatening underwater
intruder from a calculated threat level estimate based on this data.
The present invention relates to an acoustic barrier to protect an asset such as a
ship that abuts a body of water.
The present invention provides methods and apparatus for determining if an
underwater intruder passes under a protective boundary in order to protect an asset
such as a ship or a power plant. With an embodiment of the invention, a sonar sensor
system comprises a plurality of sonar sensor modules that are spaced on a protective
boundary. A sonar sensor module comprises a sonar transducer (sonar array) that is
characterized by an omni-directional radiation pattern that may overlap an omni-
directional radiation pattern of an adjacent sonar sensor module. The sonar sensor
module may receive sonar signals from reflections off a target that may be an
4
underwater intruder. The sonar sensor module collects sonar data such as range
information of the target in relation to time.
A central processor obtains the sonar data from each sonar module through a
telemetry link. The central processor processes the sonar data from the plurality of
sonar sensor modules in order to determine an estimated path of the target.
Furthermore, the central processor may determine if the target should be considered
as an underwater intruder from a threat level estimate such as a course direction, a
target motion threat score, target echo width, or a target echo amplitude.
In a variation of the embodiment of the invention, the central processor
determines the estimated path by matching sonar tracking data to different simulated
sonar tracking data, in which each simulated sonar tracking data corresponds to a
different simulated path of the target. In another variation of the embodiment, the
central processor determines an initial estimated path from geometric parameters
such as range differences and time differences that are obtained from adjacent sonar
sensor modules. The central processor adjusts the estimated path in order to minimize
an error function.
(II) Lars OSTLIE et.al., (1991) [2] An immaterial fish fence is based on a
combination of low frequency mechanical vibrations and synchronously modulated
electric fields, where fish approaching the fence, will be given at the same time fear
reactions and directional information by the mechanical vibrations, and in addition
feel pain due to the electric field. The fish will then turn and swim away. The fence is
implemented by means of columns positioned side by side, each comprising a
number of low frequency transducers suspended above each other, each column
5
being suspended in a float. Each column also has two electrical conductors to which
a high voltage can be delivered, and thus synchronized fields of both acoustic and
electric type can be generated between and around the columns.
(III) Oleg Savelievich Kochetov et.al., (2016) [3] FIELD: construction.
SUBSTANCE: fencing comprises a profiled wall and a perforated wall, between
which a layer of a sound-absorbing material is placed. One of the walls is made
smooth, and the absorbent material is arranged in two layers, one of which, a harder
one, is made continuous and profiled, and the other, a soft one, is made discontinuous
and arranged beneath the surfaces of the first layer. The fencing comprises a frame,
window and door openings, apertures to accommodate lighting fixtures,
and acoustic fencings comprising the smooth and perforated walls, between which an
absorbent material is placed, arranged in two layers, one of which, a harder one, is
made composite, consisting of alternating spherical surfaces and flat surfaces with
resonant openings, and the other, a soft one, is made discontinuous in the form of
discontinuous sound absorbers and is arranged in the focus of the sound reflecting
surfaces of the first layer. The composite layer is made of a sound-absorbing
material, whose sound reflection coefficient is greater than the sound absorption
coefficient, and the discontinuous sound absorber located in the focus of the
composite layer is made in the form of rotation bodies, such as a sphere, an ellipsoid,
a cone, a truncated cone, and is fixed on the perforated wall by means of pins, one
end of which is rigidly fixed to the perforated wall, and the other is made pointed and
arranged in the body of the discontinuous sound absorbers.
EFFECT: increasing the noise absorption efficiency.
6
CHAPTER 3
INTRODUCTION TO SOUND
Sound is a vibration that propagates as a mechanical wave of pressure and
displacement through some mediums (such as air and water).the sound waves are
generated by a sound source as the vibrating diaphragm of a speaker. The
vibrations propagate away from the source at the speed of sound, thus forming
the sound wave. Sound propagates through compressible media such as air water
and solids. During propagation, waves can be reflected, refracted, or attenuated
by the medium. The speed of the sound is depend on the medium the waves pass
through and is a fundamental property of the material n air at sea
level, he speed of sound is approximately 343 m\s.. By creating disturbance in the
air at certain frequencies we are able to create an audible noise. The frequencies
that may be audible to different beings are different. Sound is classified into
Audible and inaudible sound. Where audible is, the human hearing range (20Hz-
20kHz), and the range below 20 Hz is infrasound and above 20kHz is Ultrasound
which are inaudible. Our area of investigation is (0-20 Hz) Infrasound. These
inaudible infrasonic sound waves can be felt by humans through physical body in
the range (4-16Hz) if the amplitude is high.
7
CHAPTER 4
INFRASOUND
Infrasound is those with frequencies below human hearing, from 20
Hertz down to 0.001Hz. There are a number of human caused and natural sources.
Wind turbines, diesel generation stations and highway traffic are some common
artificial noises. Natural infrasonic sounds may be caused by weather, sandstorms,
winds, wildfires, volcanoes, waterfalls, meteors, distant tornadoes, and upper
atmosphere lightning. Infra sounds have very long wave lengths measured in
meters. These waves travel faster along the ground in all directions than sound
traveling through air. That is because the ground is much denser than air. Infra
sounds travel global distances with little loss of energy.
The belief that animals can predict earthquakes (natural disasters) has
been around for centuries. Historians recorded that animals, including rats, snakes
and weasels, deserted the Greek city of Helice in droves just days before a quake
devastated the place. This was a myth from a lot of centuries. The science behind
this myth is reveled, i.e., animals escape natural disasters only because they have
the ability to sense infrasound.
Through our study we found that the frequencies produced during
natural disasters to be in the range from 0.002Hz to 2.5Hz. Propagation of this
range through air is less effective. The same natural disaster effect can be
produced in the range of 5Hz to 40Hz in the medium of air.
8
4.1. RESONANCE
Resonance is a phenomenon in which a vibrating system or external force
drives another system to oscillate with greater amplitude at specific frequencies.
Frequencies at which the response amplitude is a relative maximum are known as
the system's resonant frequencies or resonance frequencies.
When infrasound propagates, it causes mechanical vibrations and if the
frequency of the objects is equal to that of the infrasound, the source imparts
energy to the object and both are joined and vibrate at same frequency. This
is called resonance. That means if one increases the energy of the source, energy
of the vibrating object will also increase when in resonance. In a way, one can
control the energy of the object while sitting at a place away from the object.
Organs of animal body have their natural frequency in infrasonic region.
4.2 STUDY ON ELEPHANTS AND INFRASOUND
Elephants can sense and they also use infrasound for communication,
indeed they make infrasonic calls that have powerful deep calls in long distance
communication. They make a variety of vocalizations including rumbles,
screams, and trumpets. Rumbles are low-frequency calls, often falling partially or
entirely in the infrasonic range. Most elephant rumbles consist of a fundamental
frequency between 5-30Hz with audible harmonics or overtones. With distance,
the upper harmonics attenuate at a greater rate than the lower ones. A good
working range for capturing elephant rumbles with their harmonics is 5-250 Hz.
The lowest call we have measured for forest elephants was at 5Hz; from
savannah elephants, 14Hz.
9
Circumstances in which animals emit infrasonic rumbles are, most
elephant rumbles are rich in infrasound; many of these contain frequencies high
enough to be audible to humans. Elephants make these calls when coordinating
family and larger group behaviors, when competing for resources and/or
dominance, and when attracting mates and announcing reproduction. Females
with young are most vocal. The vast majority of infrasonic calling took place in
family groups; bull groups were relatively silent. There is, however, still much to
be learned about the functions of elephant calls.
4.2.1 How often will an elephant use infrasonic calls
Females vocalize more frequently than males; the rate of vocalization in
both males and females is highly variable and dependent on social circumstances.
The number of calls per unit time increases predictably with the number of
elephants present, but the rate of calling by each elephant remains relatively
consistent.
4.2.2. Travelling of infrasonic elephant calls
The lower the frequency of a sound, the longer its sound wave, Low
frequency sounds can therefore travel farther without being absorbed or reflected
by the environment. Intensity of elephant calls varies widely from very soft calls
made between mothers and their adjacent infants to the very loud calls made by
females announcing their availability. Playback experiments demonstrated that
savannah elephants responded to each other loud vocalizations over distances of
at least 2 kilometers. Because playbacks were only broadcast at half the
amplitude of the strongest elephant calls in their sample, the authors estimated
the actual range as at least 4 kilometers.
10
The calling area may be expanded by as much as an order of magnitude
during temperature inversions in the evening and night. Preliminary results, with
forest elephants suggest that powerful forest elephant calls can be heard by
elephants 7km away through the dense forest. The fact that elephant calls can
travel several kilometers enables elephant societies to coordinate movements
over large areas.
4.2.3. Analyzing infrasonic calls
One way to discover if we have recorded infrasonic elephant calls is to
speed up the recording, raising all the frequencies in the recording to a level that
we can hear them. Typically, if we speed up a recording containing infrasonic
elephant calls 3 times, we will easily be able to hear them.
Sounds can also be represented visually using spectrograms. Spectrograms
graph frequency on the y-axis, time on the x-axis and represent loudness of sound
by the darkness of the display.
4.2.4. A unique feature of infrasonic elephant calls which distinguishes them
from other infrasound that may be recorded
The structures of elephant rumbles are quite varied, but readily
recognizable. The detection of an elephant call involves roughly a eyebrow-
shaped signal between 1-250 Hz and lasting between 2-10 seconds. A few other
infrasonic noises might be a broadband wind or thunder, which often obscures
elephant infrasound.
As our plan is to use sound waves to repel elephants from the crop fields
and hence, prevent fatalities of elephants and also the damage of crops. Elephants
have an audibility range of 12-12000 Hz and can hear infrasound. Through
11
experimental means, a particular frequency can be determined at which the
animal experiences minor irritation. A device emitting the sound of this
frequency can be placed in fields in a position sufficient enough to produce
infrasound when the elephant is passing through areas where restricted. This can
either be done manually or done automatically using a GPS or iOT which will
work along with the device. The device itself will produce the sound using
LabVIEW that we have written. Although the idea of repelling animals using
sound waves has been floating around for quite some time, what makes our
proposal unique is that we have taken into account various factors which will
affect the emitted sound’s frequency Air temperature and the speed of the train
will be taken into account and amendments to the emitted frequency will be
made.
12
CHAPTER 5
METHODOLOGY
There are various stages we came across and here we are today bringing
our idea into a product. The main motive is and was always to produce
―infrasound‖.
5.1. PRODUCTION OF INFRASOUND USING PUMPING ACTION
It has been witnessed that in jungles, carnivorous beasts like tigers and
lions are able to generate sounds at infrasound levels. Herbivorous animals are
able to sense these signals especially when these carnivorous animals are too
close to them. The generated infrasound vibrations produce serious impact over
these poor herbivorous animals – they freeze with fear and are just not able to
move an inch. This makes them like sitting ducks and they are instantly grabbed
by the deadly beasts.
The secret behind generating effective infrasound is creating a sort of
―pumping" action over an enclosed volume of air at the specified frequency To
be precise the air volume needs to be resonated at the specified frequency. This is
best done by channelizing the sound waves through pipes or narrow columns
before throwing it into the room.
The figure no 5.1 shows a circuit basically comprised of an amplifier with
the required loudspeaker set-up. The chip TDA 1521 used here consists of two
discrete amplifier modules in one package and is able to produce a good 12-watts
of audio power from each channel – quite enough for the present application.
13
Fig no 5.1 Circuit diagram of pumping action
Though it sounded to be a simple idea for generating infrasound artificially,
Application of this is quite insignificant.
Vibrations can be produced inside a closed premise only.
5.2. INFRASOUND GENERATION USING ARDUINO
Since the whole setup would completely become electrical, we went to the
next stage
14
5.3. INFRASOUND GENERATION USING CLOCK PULSE
Fig no 5.2 555 IC timer circuit
15
CALCULATIONS:
T = 0.7(R1+2R2)C1
f = 1.4/(R1+2R2)C1
Where,
T- Time period in ms.
f- Frequency in Hz.
R1 and R2- Resistance in Mohm.
Table no 5.1 Tabulations of frequencies.
Frequency, f
(Hz)
Resistance 1, R1
( x106 Ω
Resistance 2, R2
( x106 Ω
1 4.6 4.7
2 2.2 2.4
3 1.665 1.5
4 1.1 1.2
5 0.8 1.0
6 0.705 0.805
7 0.6 0.7
8 0.55 0.6
9 0.455 0.55
10 0.4 0.5
11 0.7 0.3
12 0.6 0.3
16
Frequency, f
(Hz)
Resistance 1, R1
( x106 Ω
Resistance 2, R2
( x106 Ω
13 0.5 0.3
14 0.4 0.3
15 0.35 0.3
16 0.3 0.3
17 0.25 0.3
18 0.2 0.3
19 0.15 0.3
20 0.1 0.3
21 0.2 0.22
22 0.212 0.2
23 0.202 0.2
24 0.194 0.19
25 0.186 0.18
A clock pulse produces audible sound of this infrasonic frequency range,
which should actually not happen.
5.4. PRODUCTION OF INFRASOUND USING LabVIEW
In this setup, we made use of a webcam, a set of speaker and LabVIEW as
a major platform to successfully produce sounds of variable frequencies.
17
CHAPTER 6
PROJECT BRIEF
Initially, Image processing is done. Here two kinds of images play a vital
role, the template in the Lab\VIEW and the captured image from webcam. The
templates, i.e., whenever the pre- saved image matches the image captured from
the webcam, there is a activation in the system which leads to the production of
sound of particular frequency range accordingly.
Fig no 6.1 Block diagram of LABVIEW processing
Web
camera
LabVIEW
Software
Processing
Alarm
Light
Speakers
(Infrasound)
18
6.1. IMAGE PROCESSING
6.1.1. Template creation:
Using Vision Assistant, required image is opened from Open Image tool in
the Title Bar, and from Processing Functions: Color tool, color extraction is done,
in which a green plane is extracted. Followed by Processing Function: Mission
Vision , Pattern match tool is used, to create a template on clicking New
Template there opens a dialogue box named NI Vision Template Editor- Select
Template Region, in which the unnecessary part of the image is erased to have a
clear view of what image is exactly needed for image processing is created.
6.1.2. Image from webcam:
Vision Acquisition tool is used to capture the image from the webcam and
the if it matches the image from the template then the signal is produced, which
in turn produces frequency range accordingly.
6.2. DESCRIPTION OF FRONT PANEL:
It consists of:
1. Image Out
2. Matches
3. Frequency Graph
4. Stop
5. Warning light
Image Out window is interconnected with vision acquisition tool which
acquires continuous images from the web camera. The images captured are
compared with the pre-set templates.
Matches window is interconnection with Vision assistant tool where the
whole image processing is done by comparing every coordinate points of the
19
captured and pre-set image. It gives us the data of the co-ordinates and number of
matches.
Frequency graph is interconnected with the assigned value of the structure
case. The scale of the frequency graph can be varied according to the necessity.
The frequency produced is graphically shown.
A stop switch is interconnected with the whole circuit. So, when the stop
switch is turned on, the circuit breaks and the supply is cut off. There by stopping
the production of infrasound at any instant necessary. It is a user-friendly
controlling system.
Warning light is connected with the output of the Vision Assistance tool. It
warns the people about the entry of an animal into or nearby the fields.
Fig no 6.2 Front panel
20
6.3. BLOCK DIAGRAM OF IMAGE PROCESSING.
Fig no 6.3 Bock diagram of image processing
21
6.3.1. IMAQ
IMAQ Vision for LabVIEW—a part of the Vision Development Module is a
library of LabVIEW VIs that you can use to develop machine vision and scientific
imaging applications. The Vision Development Module also includes the same
imaging functions for LabWindows and other C development environments, as well
as ActiveX controls for Visual Basic. Vision Assistant, another Vision Development
Module software product, enables you to prototype your application strategy quickly
without having to do any programming. Additionally, NI offers Vision Builder AI:
configurable machine vision software that you can use to prototype, benchmark, and
deploy applications.
6.3.2. VISION ACQUISITION
Fig no 6.4 Vision acquisition tool.
Use the Select Acquisition Source step to select the acquisition device. The
settings and options available during the Configure Acquisition Settings step vary
based on the device you select. Complete the following steps to choose a device for
acquisition:
22
1. Select a device from the list of available devices in the Acquisition Sources
control.
2. Click Acquire Single Image to acquire a single image, or click Acquire
Continuous Images to acquire continuous images. Use Zoom to Fit, Zoom
1:1, Zoom In, and Zoom Out to adjust the zoom on the acquired image.
3. Click Next.
Acquiring from a remote target:
The NI Vision Acquisition Express Wizard allows you to acquire images from
remote devices. The following steps are used to configure an acquisition on a remote
target.
1. Launch LabVIEW and create a new project.
2. Add a remote target to the project. Right-click the target. Select New»VI. This
opens a new VI and adds it under the remote target to the project.
3. Right-click the block diagram of the new VI to display the Functions Palette.
Select Vision and Motion»Vision Express»Vision Acquisition Express VI
and drag it to the block diagram. This launches the NI Vision Acquisition
Express Wizard.
4. All the devices connected to the remote target appear in the list of devices in
the Acquisition Sources control on the NI Vision Acquisition Express
Wizard.
5. Select the device from the list of available devices in the Acquisition Sources
control.
6. Click Next.
23
6.3.3. Vision assistant
Fig no 6.5 Vision assistant tool
Triggers tab
Use the Triggers tab to synchronize an image acquisition with events
external to the computer, such as receiving a pulse from a sensor to indicate the
position of an item on an assembly line. The trigger signal must be connected to a
trigger input of the image acquisition device. The triggered acquisition settings
available in this step vary based on the type of device you select in the Select
Acquisition Source step and the acquisition type you select in the Select Acquisition
Type step. The Trigger Summary Table provides an overview of all the triggers
configured for this acquisition.
The following steps are used to configure a trigger line.
1. Click Add Trigger to add a trigger to the Trigger Summary Table.Click
Remove Trigger to remove a trigger from the Trigger Summary Table.
2. Select the Line to use for the trigger signal.
24
3. Use the Action control to specify the trigger type of the selected Line. The
following options are available:
o Trigger Start of Acquisition—An acquisition begins when the specified
trigger line detects the correct polarity edge.
o Trigger Every Image—Waits until the specified trigger line detects the
correct polarity edge to acquire each image.
o Trigger Start of Buffer List—Acquires the buffer list after receiving the
assertion edge of a trigger. The first buffer fills on trigger signal. This
mode is only supported for the Continuous Acquisition with Inline
Processing with the Acquire Image Type set to Acquire Every Image.
o Trigger Every Line—Acquires each line from a line scan camera after
the specified number of Skip Triggers edges occurs. This mode only
applies to line scan cameras.
o Trigger End of Acquisition—Stops the acquisition after acquiring the
Number of Post Trigger Buffers. Returns an array of pre-trigger images
and post-trigger images. This mode is only supported for the Continuous
Acquisition with Inline Processing with the Acquire Image Type set to
Acquire Every Image.
o Trigger Height of Acquisition—Uses the trigger signal level to
determine how long to acquire from a line scan camera. Each line
acquired during an active trigger line is added to the image height. When
the trigger line becomes inactive, the acquisition results return. This
mode only applies to line scan cameras and cannot be used at the same
time as Trigger Every Image.
4. Select a Polarity to define the active edge of the trigger. The following options
are available:
25
o Rising Edge—Detects light edges against a dark background.
o Falling Edge—Detects dark edges against a light background.
5. Specify a Timeout, in milliseconds, to wait for a trigger to receive an active
edge and capture an image.
6.3.4. BLOCK DIAGRAM INPUTS
Parameter And Description
Image Src - Is a reference to the source image.
Image Dst - Is a reference to the destination image. If Image Dst is connected, it must
be the same type as the Image Src.
error in (no error) - Describes the error status before this VI or function runs. The
default is no error. If an error occurred before this VI or function runs, the VI or
function passes the error in value to error out. This VI or function runs normally only
if no error occurred before this VI or function runs. If an error occurs while this VI or
function runs, it runs normally and sets its own error status in error out. Use the
Simple Error Handler or General Error Handler VIs to display the description of the
error code. Use error in and error out to check errors and to specify execution order
by wiring error out from one node to error in of the next node.
26
6.3.5. Block diagram outputs
Image Dst Out
Is a reference to the destination image. If Image Dst Out is connected, Image Dst
Out is the same as Image Dst. Otherwise, Image Dst Out refers to the image
referenced by Image Src.
Error out
Contains error information. If error in indicates that an error occurred before this
VI or function ran, error out contains the same error information. Otherwise, it
describes the error status that this VI or function produces. Right-click the error
out indicator on the front panel and select Explain Error from the shortcut menu
for more information about the error.Configuring Image Logging Settings
The following steps are used to enable image logging.
1. Select the Enable Image Logging checkbox.
2. Click Browse to select the folder to save images.
3. In the File Prefix text box, enter the text to use for the filename prefix. The default
text is Image.
4. In the File Format dropdown listbox, select a file format to save the files. You can
select BMP, TIFF, JPEG, JPEG2000, PNG, or AVI. The file formats have
different options under File Format Settings.
27
Fig no 6.6 Block diagram (1)
Fig no 6.7 Block diagram (2)
28
CHAPTER 7
CONCLUSION
Through our study and experiment we found that animals can detect
natural disasters even before humans or other instruments can, as they have the
ability to sense and hear infrasound. By generating infrasound only after sensing
the animal near the crop field, our project comes into play and the sounds
produced are effective, in restricting their entry into fields. Hereby a two way
advantage is achieved, where the crops and the lives of animals are given equal
importance for a better living on earth.
29
CHAPTER 8
COST ESTIMATION
Table no 8.1
NO MATERIALS QUANTITY COST
1 Web camera 1 ₹ 725
2 Speaker 1 ₹395
3 DAQ 1 ₹7000
TOTAL ₹8120
30