<?xml version="1.0" encoding="UTF-8"?>
<record
    xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
    xsi:schemaLocation="http://www.loc.gov/MARC21/slim http://www.loc.gov/standards/marcxml/schema/MARC21slim.xsd"
    xmlns="http://www.loc.gov/MARC21/slim">

  <leader>04683nam a22001697a 4500</leader>
  <controlfield tag="003">NUST</controlfield>
  <datafield tag="082" ind1=" " ind2=" ">
    <subfield code="a">005.1,KAN</subfield>
  </datafield>
  <datafield tag="100" ind1=" " ind2=" ">
    <subfield code="a">Kanwal, Maria </subfield>
    <subfield code="9">35176</subfield>
  </datafield>
  <datafield tag="245" ind1=" " ind2=" ">
    <subfield code="a">Development of Improved Saliency Detection Techniques /</subfield>
    <subfield code="c">Maria Kanwal</subfield>
  </datafield>
  <datafield tag="260" ind1=" " ind2=" ">
    <subfield code="a">Rawalpindi,</subfield>
    <subfield code="b">MCS (NUST),</subfield>
    <subfield code="c">2022</subfield>
  </datafield>
  <datafield tag="300" ind1=" " ind2=" ">
    <subfield code="a">xviii, 134 p</subfield>
  </datafield>
  <datafield tag="505" ind1=" " ind2=" ">
    <subfield code="a">Saliency is the perceptual quality by which an observable thing or object stands out relative
to its environment. It is that property by which some object stands out. Accurate detection
of salient object is a challenging task. In this thesis, novel and robust saliency detection
techniques and their different applications are proposed to detect salient objects from images.
Different features of images are utilized to extract accurate saliency maps.
The first proposed technique formalizes the saliency map using color and depth along
with input&#x2019;s histogram. The color saliency map assumes that the salient objects in an image
are usually more colorful than objects that are not salient. Likewise, the salient objects are
usually found closer to the camera during acquisition; therefore, the transmission map is also
utilized for saliency detection. The third map is estimated from the histogram. These three
maps are then fused and filtered in order to yield a smooth saliency map.
The second proposed technique is based on texture contrast and convex hull center prior
maps. Input image is first segmented into super-pixels. Primary saliency map is constructed
using texture features. Convex-hull-based center prior map is extracted and refined with
graph regularization to obtain secondary map. Primary and secondary maps are fused based
on weight matrix. Thresholding and filtering are applied to obtain final map.
The third proposed technique is frequency and gradient domain saliency detection using
pyramid approach. The edges of the input image are sharpened first. Pyramid decomposition
at different scales is applied. At each scale, the color space of the RGB image is converted
into YIQ to extract the Q channel. In the frequency domain, Butterworth high-pass filtering
is applied. Gradient and DoG filters are applied. Each channel&#x2019;s saliency map is created via
region filling. The saliency maps of pyramids at different scales are integrated and filled.
The fourth proposed saliency detection technique is for underwater images. The input
image is enhanced and gradient domain filtering is applied. The filtered image undergoes
a pyramid decomposition. Saliency is computed on three different scales on two parallel
steps. In first step, saliency is computed using cellular automata at each scale. The map is
further optimized and filtered to obtain saliency at a given scale. A primary saliency map is
computed by scale integration. In second step, saliency is computed using a homomorphic
filter on each scale. Guided filter is applied to obtain saliency at a given scale. Secondary
saliency map is computed by scale integration. Multiplicative fusion is applied on primary
and secondary saliency maps to obtain final output.
In the fifth proposed technique, saliency based fabric defect detection via bag-of-words
iv
model is proposed. The proposed methodology feeds the fabric images into the framework.
The proposed saliency technique is used to create saliency maps of fabric images. The kmean
clustering algorithm is used to extract local features. A bag of features is created
using cluster centers. To detect defected fabric, feature vectors are used to train a K-Nearest
Neighbor (K-NN) classifier.
In the sixth proposed technique, it is proposed to direct the user&#x2019;s attention to a specific
region of interest (ROI). To begin with, a bilateral filter is used to smooth ROI by removing
unwanted details from the defocus map. After that, the hue component of the ROI is adjusted,
and the outline is extracted using neighboring data. Lastly, the hue component is used to
direct the viewer&#x2019;s attention to a specific region in the resultant image&#x2019;s outline.
Cutting-edge existing techniques are used to compare the proposed techniques. Experiments
demonstrate the significance of proposed techniques on various datasets. The simulation
results show that the proposed techniques produce better results that are noise-free
and very close to the ground truth. These salient detection techniques can be used in various
applications like image and video compression, object tracking, image segmentation, etc.</subfield>
  </datafield>
  <datafield tag="650" ind1=" " ind2=" ">
    <subfield code="a">PhD Computer Software Engineering Thesis</subfield>
    <subfield code="9">132801</subfield>
  </datafield>
  <datafield tag="651" ind1=" " ind2=" ">
    <subfield code="a">PhD CSE Thesis</subfield>
    <subfield code="9">132802</subfield>
  </datafield>
  <datafield tag="700" ind1=" " ind2=" ">
    <subfield code="a">Supervised by Dr. Abdul Ghafoor</subfield>
    <subfield code="9">132894</subfield>
  </datafield>
  <datafield tag="942" ind1=" " ind2=" ">
    <subfield code="2">ddc</subfield>
    <subfield code="c">THE</subfield>
  </datafield>
  <datafield tag="999" ind1=" " ind2=" ">
    <subfield code="c">615919</subfield>
    <subfield code="d">615919</subfield>
  </datafield>
  <datafield tag="952" ind1=" " ind2=" ">
    <subfield code="0">0</subfield>
    <subfield code="1">0</subfield>
    <subfield code="2">ddc</subfield>
    <subfield code="4">0</subfield>
    <subfield code="7">0</subfield>
    <subfield code="a">MCS</subfield>
    <subfield code="b">MCS</subfield>
    <subfield code="c">THE</subfield>
    <subfield code="d">2026-01-26</subfield>
    <subfield code="l">0</subfield>
    <subfield code="o">005.1,KAN</subfield>
    <subfield code="p">MCSPhD CSE-15</subfield>
    <subfield code="r">2026-01-26</subfield>
    <subfield code="w">2026-01-26</subfield>
    <subfield code="y">THE</subfield>
    <subfield code="z">Almirah No.68, Shelf No.5</subfield>
  </datafield>
</record>
