Skip to main content

Comparative study: enhancing legibility of ancient Indian script images from diverse stone background structures using 34 different pre-processing methods

Abstract

In recent times, there has been a proactive effort by various institutions and organizations to preserve historic manuscripts as repositories of traditional knowledge and cultural heritage. Leveraging digital media and emerging technologies has proven to be an efficient way to safeguard these invaluable documents. Such technologies not only facilitate the extraction of knowledge from historic manuscripts but also hold promise for global applications. However, transforming inscribed stone artifacts into binary formats presents significant challenges due to angle distortion, subtle differences between foreground and background, background noise, variations in text size, and related issues. A pivotal aspect of effective image processing in preserving the rich information and wisdom encoded in stone inscriptions lies in employing appropriate pre-processing methods and techniques. This research paper places a special focus on elucidating various preprocessing techniques, encompassing resizing, grayscale conversion, enhancement of brightness and contrast, smoothening, noise removal, morphological operations, and thresholding. To comprehensively assess these techniques, we undertake a study involving stone inscription images extracted from the Tanjore Brihadeeswar Temple, dating back to the eleventh century during the reign of Raja Raja Chola. This choice is informed by the manifold challenges associated with image correction, such as distortion and blurring. We undertake an evaluation encompassing a diverse array of stone background structures, including types like flawless-bright-moderately legible, dark-illegible, flawless-bright-illegible, flawless-dull, flawless-irregular-moderate, highly impaired-dark-legible, highly impaired-irregular-illegible, impaired-dark-moderate, impaired-dull-moderately legible, impaired-dusky dark-moderate, and very impaired-dusky dark-legible. Subsequently, the processed outputs are subjected to character recognition and information extraction, with a focus on comparing the outcomes of various pre-processing methods, including binarization and grayscale conversion. This study seeks to contribute insights into the most effective pre-processing strategies for enhancing the legibility and preservation of ancient Indian script images etched onto diverse stone background structures.

Introduction

The ancient technologies of bygone eras are often regarded as advanced and distinct from the technologies we employ today. One remarkable example is the Brihadeeswara Temple, constructed in the eleventh century. The engineering marvel of erecting such a colossal temple remains a mystery, given its massive size of 216 feet and the use of interlocking stones without any apparent cement or bonding material. To this day, constructing a building or structure of similar magnitude without modern cementing materials remains an unparalleled feat. These structures, along with other artifacts such as inscribed stones, palm leaves, copper plates, wooden artifacts, pillars, temple walls, rock beds, and potsherds, are believed to hold valuable information and knowledge from ancient times. The primary purpose of such inscriptions and repositories is to safeguard and transmit historical information across generations [1,2,3].

The advent of image processing, data science, machine learning, and artificial intelligence presents a unique opportunity for the modern world to decipher and analyze ancient information. A substantial volume of Tamil historical documents, once stored in libraries, museums, and temples, is now being digitized and made accessible through digital platforms [4,5,6]. However, this digitization process is met with numerous challenges. Many images and characters are in various stages of degradation and distortion, marred by noise and disruptions. Yet, the extraction of invaluable information from these artifacts can significantly enrich our understanding and be applied to establish literacy, archaeology, and historical context. Ancient Tamil inscriptions can be categorized into three main types: Vatteluthu, Grantha, and Tamil Brahmi. The Brahmi script, originating in the Ashoka era, served as the precursor to nearly all Indian scripts, including Vatteluthu and Grantha.

The pristine Brihadeshwara temple in Tanjore, constructed between 1003 and 1010 AD by Raja Raja Chola I, stands as a repository of valuable information and Vedic technological knowledge. However, the focus of this paper lies in deciphering scripts that have endured varying degrees of decay and degradation, compounded by a multitude of noises. The central objective of preprocessing techniques is to enhance the image quality by effectively mitigating undesirable distortions while enhancing crucial image features, thus preparing the images for subsequent processing steps. Despite employing advanced photography protocols and sophisticated camera and scanning equipment, historical inscription images often remain unreadable due to the inexorable ravages of time. The deterioration of inscriptions can be attributed to factors such as ink migration, cracking, damages from biological and environmental influences, as well as foreign elements like dirt and discoloration (Fig. 1).

Fig. 1
figure 1

Brihadeshwara temple of Tanjore Inscription sample and century wise ancient character representation

Addressing this challenge entails a series of steps (Fig. 2). The first step involves converting images to grayscale and resizing them for uniformity. The second step encompasses enhancing brightness and contrast using histogram analysis and alpha–beta comma values. The third step centers on smoothing and noise reduction, accomplished through filtering techniques like Gaussian, median, and bilateral filters. Subsequently, thresholding techniques are employed to distinguish between dark and light regions, utilizing methods such as adaptive thresholding, OSTU, Niblack, and Savuola [7, 8]. The final step integrates morphological processing to enhance characters, particularly in cases where thresholding or other preprocessing steps may have eroded or added pixels to characters. There are many languages in India and so the diversity in scripts throughout, much research has been documented about preprocessing of ancient inscriptions; however holistic approaches to understanding these recent technologies and evaluation of all such techniques are very few [9]. For instance; Buzykanov [10] proposed the methodology for improving pixel density of text images via low pass signal Gaussian Laplacian filtering algorithm; however, evaluation of such technique and comparing this method with other efficient methods are not been documented since then. Similarly, an adaptive binarization method proposed by Kavel-lieratou et al. [11]; serve as an iterative mix of international and local thresholding. This method suggests the use of adjacent pixels to measure the average pixel value from the text to distinguish between text and nontext zones. Parul Saber and Sunjay proposed the multilingual character segmentation and recognition schemes for Indian ancient document images; which involve fewer sub-processes like binarisation, resizing, skew correction, and thinning to improve the clarity of images. The proposed methodology uses Otsu’s binarization technique which converts grey scaling concerning image size correction and pixel resolution to a binary image. Such methods imply the importance of backgrounds on the inscriptions; Seeger and Dance [12] estimate the intensity of background region and process binarization as compared to threshold intensity. Their algorithm calculates non-text region intensities at each pixel by which an appropriate threshold surface can be calculated [13]. In such context and advancement of the algorithm to evaluate the image text; historical handwritten images are also subjected to pre-processing; Lakshmi and Patvardhan [14] proposed pre-processing and classification methodology for Telugu handwritten characters. Their research emphasizes gathering the dataset of handwritten ancient Telugu characters from numerous scripts and printed on high-quality paper [14]. The number of basic handwritten Telugu characters considered in this proposed system is 50 and their system considered 18,000 samples in total (50 × 360). All the documents are collected from various scribers and they are scanned at 300 dpi and stored as digital images.

Fig. 2
figure 2

The sample dataset image undergoes a sequential image pre-processing pipeline

The intensity and threshold of an image are crucial in character recognition, recent research suggested projection-based text line segmentation with a variable threshold [15]. To enhance the quality of images, in preprocessing they suggested converting the colour image to a grey scale image, further binarization, and then noise reduction. Greyscale conversion method, images are evaluated by calculating the weight of Red, Green, and Blue (primary colours) components from each colour pixel, and further grey images are converted to binary images using the threshold processing method. Threshold values in binary images were calculated using Otsu’s Binarization method. And morphological errors in the background are removed by employing the morphological opening and closing operation.

The use of greyscale conversion and detection with skew angle; inscription images input RGB conversion and further smoothing on greyscale images would be an efficient method to reduce the high-frequency noise using Wiener filter. Finally, the resulted grey scale pixel values can have used in detecting skew angle. In addition, text position differences in the image were compared to evaluate whether the document is left-skewed or right-skewed.

Using the various method, the methods are Skew detection and correction, Binarization, noise removal, and morphological operations. These are techniques used in the proposed system. Panyama et al. studied the palm leaf character recognition system using a transform-based technique; for which digitalization and storage of palm leaf manuscripts utilizes a specific 3D function which is proportional to the pressure exerted at that stage by the scriber [16]. The advantage is that in the YZ plane, the precision is higher at 96 per cent and the demerit is smaller than in other planes as reported.

Ptak et al. studied and reported projection-based text line segmentation with variable threshold for which new algorithm were created for text line separation in handwriting using the projection profile. It employs thresholding, but the threshold value is variable [15]. This permits the determination of low overlapping peaks of the graph. The disadvantage of this paper is that the algorithm does not deal well with slanting and curved text lines.

This research work is devoted to presenting an experimental methodology focused on enhancing the inscriptions from the Brihadeeswara Temple in Tamil stone through an array of preprocessing methods. The intent is to unlock and revive the valuable historical information embedded in these artifacts.

Research aim

The aim of this research is to enhance the readability and extract valuable historical and technological information from ancient inscriptions, focusing on the Brihadeeswara Temple in Tanjore. These inscriptions hold crucial insights into advanced Vedic building technologies and historical narratives. The objective is to employ modern digital technologies, including image processing, data science, machine learning, and artificial intelligence, to overcome the challenges posed by the degradation, distortion, and noise present in the aged inscriptions. The primary focus will be on inscriptions in various states of decay and with significant levels of noise. The research will involve a multi-step preprocessing approach to restore the image quality, rectify distortions, and enhance critical image features. This includes techniques such as grey conversion, resizing, brightness and contrast correction using histogram adjustments, and filtering methods like Gaussian, median, and bilateral filters for noise reduction. The application of adaptive thresholding and morphological processing will also be explored to separate dark and light regions and improve character visibility. By employing these advanced techniques, the study aims to uncover the hidden knowledge encoded in the inscriptions, contributing to a deeper understanding of Vedic technologies and historical contexts. The successful extraction of information from these inscriptions could potentially provide insights into architectural practices and ancient building methods.

Methodology

Dataset

This paper takes a camera-captured Stone inscription as the Input. The inscriptions in Tanjore Brahadeswar Temple, are shown in Figure. These data are different in size and background. The camera used to capture these images are of very high resolution or quality (DSLR). The default format of these images is png or jpeg. These data cannot be used directly for the training purpose since it belongs to several years back, the letter impressions were faded. Hence Archeology department uses white chalks or paints to distinguish the characters and background. Character recognition for this language imposes a challenge as there is a high range of noise detected in stone images. Hence Preprocessing is the foremost step applied to stone images. Figure 1 camera captured original image from Archeology record.

Preprocessing of images

Image enhancement includes mechanisms for enhancing image quality, allowing improved visual and computational analysis. It is commonly used in many applications because of its ability to solve some of the limitations that image acquisition systems pose. Deblurring, removal of noise, and improving contrast are some examples of image enhancement operations. Some of the image quality factors are discussed below.

Thresholding

One of the simplest, most powerful, and most frequently used segmentation algorithms is thresholding-based segmentation. It is useful in discriminating the foreground from the background. Thresholding yields a binary image, which reduces the complexity of data and simplifies the process of recognition and classification [17, 18]. There are three types of thresholding approaches, namely, Global, Local, and Adaptive.

The adaptive mean is the basic method of thresholding in which the current pixel value of the image is replaced with mean or average of all the neighboring pixels and that value is compared with current pixel value. If the value of current pixel is less than the mean value then it is set to black otherwise it is set to white. Adaptive thresholding using a Gaussian filter depends on value of standard deviation. As the value of standard deviation increases more noise is suppressed but the image also gets blurred respectively.

A threshold Flm is a value such that

$$O (i,j )=\left\{\left\{\begin{array}{ll}0,\quad & lm<Flm*\frac{(1-t)}{100} \\ 255, \quad & otherwise\end{array}\right.\right.$$

where O(i, j) is the binarized image and lm  [0,1] be the intensity of a pixel at location (i, j) of the image.

A global thresholding technique makes use of a single threshold value for the whole image, whereas a local thresholding technique makes use of unique threshold values for the partitioned sub-images obtained from the whole image. In adaptive thresholding, for each pixel in the image, a threshold has to be calculated. However, automatic selection of optimally significant robust values is a difficult challenge. If the pixel value is below the threshold, it is set to the background value; otherwise, it assumes the foreground value [19, 20].

Otsu thresholding method [21] is a highly applicable fully automatic global thresholding algorithm. It is based on separability maximization in the grey level classes. Otsu's method looks for the threshold value which can minimize the intra-class variance. The intra-class variance can be defined as the two classes’ weighted sum of variances.

$$g\left(i,j\right)=\left\{\begin{array}{cc}1& iff\left(i,j\right)>T\\ 0& iff\left(i,j\right)\le T\end{array}\right.$$

Assuming an image is represented in L {0, 1, 2 ,…, L−1} gray levels, the number of pixels at level i is denoted by \({{\varvec{n}}}_{{\varvec{i}}}\), and the total number of pixels is denoted by i \(MN={n}_{0+}{n}_{1+}{n}_{2+}\dots .{n}_{L-1}\).

select a threshold \(T\left(k\right)=k,0<k<L-\), and use it to.

Assume the image is divided into two categories C1 and C2 (target and background) with threshold k, then C1 and C2 respectively correspond to the pixels whose grey levels are {0,1,…, k} and { k + 1,k + 2,…, L−1}.

The gray level probability distributions for the two classes are \({p}_{1}\left(k\right)=\sum_{i=0}^{k}{p}_{i}\), \({p}_{2}\left(k\right)=\sum_{i=k+1}^{L-1}{p}_{i}=1-{p}_{1}\left(k\right),{p}_{1}{m}_{1}+{p}_{2}{m}_{2}+{m}_{G}\)

\({p}_{1}+{p}_{2}=1\). The global variances for the two classes are \(\sigma \genfrac{}{}{0pt}{}{2}{G}=\sum_{i=0}^{L-1}{\left({i-m}_{G}\right)}^{2}{p}_{i}\)

The class variance and measure separabiltiy for the two classes are \(\sigma \genfrac{}{}{0pt}{}{2}{B}(k)={p}_{1}{p}_{\begin{array}{c}2\end{array}}{\left({{m}_{1}-m}_{2}\right)}^{2}=\frac{{\left(\left({m}_{G}{p}_{1}\left(k\right)-{m}\left(k\right)\right)\right)}^{2}}{{p}_{1}\left(k\right)(1-{p}_{1}\left(k\right))}\), \(\eta =\frac{\sigma \genfrac{}{}{0pt}{}{2}{B}}{\sigma \genfrac{}{}{0pt}{}{2}{G}}\)

In the gray range [0, L], the t with maximum \(0\le \eta ({k}^{*})\le 1\) is the optimal threshold.

Niblack and savoula

Niblack proposed a simple local adaptive threshold, where a threshold [22,23,24,25] is determined for each pixel based on statistics computed from a local window centered on the pixel of interest [26]. Because the threshold is adaptive, it can potentially handle cases of foreground and background intensity distribution overlap (e.g., compare Figs. 2 to 3). Specifically, Niblack thresholding uses the local mean and local standard deviation:

$$\mu \left(i,j\right)=\frac{1}{{w}^{2}}\sum_{{i}{\prime}=i-w}^{i+w}\sum_{{ij}{\prime}=j-w}^{j+w}I\left({i}{\prime},{j}{\prime}\right)$$
$$\eta = \frac{{\sigma_{{\text{B}}}^{2} (k)}}{{\sigma_{{\text{G}}}^{2} }}$$
$$\left(i,j\right)=\sqrt{\sum_{{i}^{\mathrm{^{\prime}}}=i-w}^{i+w}\frac{\sum_{{ij}^{\mathrm{^{\prime}}}=j-w}^{j+w}{\left(I\left({i}^{\mathrm{^{\prime}}},{j}^{\mathrm{^{\prime}}}\right)-\mu \left(i,j\right)\right)}^{2}}{{w}^{2}}}$$
$${\text{O}}\, \le \eta (k*) \le 1$$

where w is called the window size and controls how much context is used to compute these statistics. The per-pixel Niblack threshold is then

$${T}_{N}\left(i,j\right)=\mu \left(i,j\right)+k\sigma \left(i,j\right)$$

where k is a user-set parameter that controls the trade-of between foreground detection precision and recall. The recommended parameter setting is k =  − 0.2, though the optimal k depends on the image and chosen window size. Binarization is then accomplished with

Fig. 3
figure 3

Sample images of stone inscriptions with diverse backgrounds

$$B\left(i,j\right)=\left\{\begin{array}{cc}0& I\left(i,j\right)<{T}_{N}\left(i,j\right)\\ 255& I\left(i,j\right)\ge {T}_{N}\left(i,j\right)\end{array}\right.$$

One issue with Niblack is when the window covers only background pixels, it causes the darkest background pixels to be set to foreground (Fig. 3). While this noise is often large, the background immediately around the text is correctly identified, which makes Niblack thresholding useful in combination with other binarization techniques.

Sauvola Sauvola and Pietikäinen proposed a variant of Niblack to solve the problem with background-only windows.

$${T}_{S}\left(i,j\right)=\mu \left(i,j\right)\left[1+k\left(\begin{array}{c}\frac{\sigma \left(i,j\right)}{R}\end{array}-1\right)\right]$$

where \(\mu \left(i,j\right)\) and \(\sigma \left(i,j\right)\) are computed as in Niblack, k = 0.5 is the recommended value for the user-set parameter, and R is a constant set to the maximum possible standard deviation, i.e., R = 128 for 256 Gy levels. While Niblack takes \(\mu \left(i,j\right)\) and adjusts downward based only on the \(\sigma \left(i,j\right)\) Sauvola adjusts downward based on\(\mu \left(i,j\right)\sigma \left(i,j\right)\). In windows of only background, \(\mu \left(i,j\right)\) is relatively large, so TS < TN, which means fewer of these background pixels are set to foreground.

In analyzing heritage stone inscriptions, selecting a thresholding method hinges on diverse factors: inscription traits, image quality, noise levels, and specific analysis goals. Global thresholding, exemplified by Otsu's Method, is effective with uniform lighting but falters in varied conditions, struggling to establish a single threshold for the entire image. On the contrary, local adaptive methods like Niblack and Sauvola excel in handling lighting variations and background textures, adjusting thresholds locally for distinct inscription areas with varying contrast or degradation. While they demonstrate adaptability, precise parameter tuning—like window size (w) and user-defined parameters (such as ‘k’ in Niblack or ‘R’ in Sauvola)—is essential. These local adaptive methods, owing to their capability to adapt thresholds based on local statistics, can be particularly beneficial in deciphering text from backgrounds in challenging scenarios where lighting or degradation varies across the inscription. Nonetheless, achieving optimal segmentation necessitates careful parameter experimentation and potential combination with other techniques to mitigate specific limitations encountered with individual thresholding methods.

Blur

Median blur

The median filter [27] is a nonlinear signal processing technology based on statistics. The noisy value of the digital image or the sequence is replaced by the median value of the neighborhood (mask). The pixels of the mask are ranked in the order of their grey levels, and the median value of the group is stored to replace the noisy value.

The median filtering output is

$$y[m,n]=meidan\left\{x\left[i,j\right],|(i,j)\epsilon \omega \right\}$$

where x[i,j],y[m,n] are the original image and the output image respectively, W is the two-dimensional mask: the mask size is n*n(where n is commonly odd) such as 3*3, 5*5, etc.; the mask shape may be linear, square, circular, cross, etc.

Gaussian blur

The implementation of a 2-D Gaussian filter is widely used for smoothing and noise removal. It requires lots of computational resources and its efficiency in the implementation has been a motivating research area. Convolution operators are the Gaussian operators and the idea of Gaussian smoothing is achieved by convolution (Hypermedia Image Processing Reference 1994). The Gaussian operator in 1-D is given as:

$${G\left(x\right)=1/\sigma \sqrt{2\pi }}^{{e}^{-\frac{{(x-a)}^{2}}{2{\sigma }^{2}}}}$$

The Gaussian operator in 2D (circularly symmetric) is given as:

$${G\left(x,y\right)=1/2\pi {\sigma }^{2}}^{{e}^{-\frac{{x}^{2+}{y}^{2}}{2{\sigma }^{2}}}}$$

where σ (Sigma) indicates the standard deviation of the Gaussian function. If it has a large value, the image smoothing effect will be higher. (x, y) indicates the Cartesian coordinates of the image which show the dimensions of the window.

This filter is composed of addition and multiplication processes between the image and the kernel, where the image is represented by a matrix with a value from 0 to 255 (8 bits). The kernel is a normalized square matrix (values between 0 and 1). The kernel is represented by several bits. For the convolution process, the product of each bit of the kernel and each element of the image is then divided by a power of 2.

In the preprocessing of stone inscriptions, the median filter serves a valuable role in reducing sudden, unwanted noise, often caused by scratches or imperfections on the stone surface. It effectively replaces noisy pixel values with the median value of their neighborhood, contributing to a clearer depiction of the inscribed content. On the other hand, the Gaussian blur is adept at smoothing out minor irregularities and noise present in the image. Importantly, it achieves this without compromising the fundamental integrity of the inscribed text or intricate details. Together, these preprocessing techniques work harmoniously to enhance the overall quality and legibility of stone inscriptions for subsequent analysis or interpretation.

Enhance brightness and contrasts

CLAHE’s(Contrast-limited adaptive histogram equalization) [28, 29] basic idea is to perform the histogram equalization of the image's non-overlapping sub-areas, using interpolation to correct boundary inconsistencies. CLAHE also has two important hyper parameters: Clip Limit (CL) and Tile Number (NT). The first one (CL) is a numeric value that governs the noise amplification. When the histogram of and sub-area is determined, they are redistributed in such a way that its height does not surpass a specified “clip limit.” Instead, the total histogram is determined to perform the equalization. The second (NT) is an integer value that governs the sum of non-overlapping sub-areas: the image is divided into many (usually squared) non-overlapping regions of similar sizes, based on its value [30,31,32,33].

If numbers of pixels and grayscales, in each region, are respectively M and N, and if hi,j(n), for n = 0, 1, 2,…, N−1, is the histogram of (i, j) region, then an estimate of the corresponding CDF, properly scaled by (N−1) for grayscale mapping, is

$${f}_{i,j}\left(n\right)=\frac{N-1}{M}.\sum_{k=0}^{n}{h}_{i,j}\left(k\right);$$
$$n=\mathrm{1,2},3\dots .N-1$$

This function can be used to convert the given grayscale density function, approximately, to a uniform density function. This procedure is referred to as histogram equalization. In order to limit the contrast to a desired level, the maximum slope of (1) is limited to a desired maximum slope. One approach in limiting the maximum slope is to use a clip limit β to clip all histograms. The clip limit β is obtained by:

$$\beta =\frac{M}{N}\left(1+\frac{\alpha }{100}\left({s}_{max}-1\right)\right)$$

where α is a clip factor, if clip factor is equal to zero the clip limit becomes exactly equal to (M/N) results into an identity mapping by evenly distributing all regional pixels into all possible grayscales. Moreover, if clip limit is equal to 100 the maximum allowable slope is smax.

Gamma, alpha and beta correction: Gamma correction [34,35,36] is a non-linear adjustment to individual pixel values. While in image normalization we carried out linear operations on individual pixels, such as scalar multiplication and addition/subtraction, gamma correction carries out a non-linear operation on the source image pixels, and can cause saturation of the image being altered.

$${I}_{out}=c{{I}_{in}}^{\Upsilon}$$

where \({I}_{in}\) and \({I}_{out}\) are the input and output image intensities, respectively. c and Υ are two parameters that control the shape of the transformation curve. In gamma correction, controls the slope of the transformation function. The higher the value of Υ is, the steeper the transformation curve becomes. And the steeper the curve is, the more the corresponding intensities are spread, causing more increase of contrast [37, 38].

Adjusting the brightness mean, either increasing the pixel value evenly across all channels for the entire image to increase the brightness, or decreasing the pixel value evenly across all channels for the entire image to decrease the brightness.

$$g\left(x)\ =\, \propto f(x)\right.+\beta$$

The parameters α > 0 and β are often called the gain and bias parameters; sometimes these parameters are said to control contrast and brightness respectively [39, 40].

These methods, particularly CLAHE and gamma correction, when applied to stone inscriptions, work synergistically to optimize contrast, control noise, and enhance the overall quality of the inscribed content. By fine-tuning parameters and employing adaptive techniques, these methods cater specifically to the challenges posed by stone inscriptions, resulting in improved readability and preservation of historical content.

Bilateral filter

Bilateral filtering [41] is a non-linear filtering method, where the weight of each pixel is determined using a Gaussian in the spatial domain, multiplied by an impact function in the intensity domain, which decreases the weight of pixels with large variations in intensities. Pixels which vary greatly in intensity from the central pixel are weighted less, even though they may be near to the central pixel. It is then implemented as two Gaussian filters in a localized pixel neighborhood, one in the space domain, called the domain filter that smoothes homogeneous regions, and one in the strength domain, called the range filter that regulates edge preservation smoothing. The key advantage of using bilateral filters is therefore the development of broad and homogeneous areas [42].

$$g\left(x\right)=\left(f*{G}^{s}\right)\left(x\right)=\underset{R}{\int }f\left(y\right){G}^{s}\left(x-y\right)dy$$

The weight for f(y) is equal to \({G}^{s}\left(x-y\right)\) and depends only on the distance from space x − y. The bilateral filter introduces a concept of weighting that is dependent on the tonal distance f(y)−f(x). The result:

$$g\left(x\right)=\frac{\underset{R}{\int }f\left(y\right){G}^{s}\left(x-y\right){G}^{t}\left(f\left(x\right)-f\left(y\right)\right)dy}{\underset{R}{\int }{G}^{s}\left(x-y\right){G}^{t}\left(f\left(x\right)-f\left(y\right)\right)dy}$$

Remember that since the weights are directly dependent on the values of the image, we need explicit normalization so that the 'sum' of all weights equals one [43].

When applied to stone inscriptions, bilateral filtering proves invaluable in refining image quality by smoothing homogeneous regions while maintaining the integrity of intricate details and edges within the inscriptions. Its ability to selectively smooth and preserve edges makes it a powerful tool for enhancing the clarity and readability of the inscribed content.

Resize and gray scale

Images with a large size also have to be resized to a smaller size that is adequate for appropriate distinguishing, because an increase in the size of the input image results in an increase in the parameter to be measured, the computing power needed, and memory. Gray scale conversion process removes all color information, leaving only the luminance of each pixel [43, 44].

In the context of stone inscriptions, these methods—resizing large images and converting to grayscale—are crucial for optimizing computational resources, memory usage, and focusing on the inherent content details essential for analysis and interpretation. They ensure efficient processing while retaining critical information necessary for further analysis.

Dilation and erosion

Dilation is a morphological operation which extends the image to bright structures. To this end, the new gray-value of each pixel in the structuring factor centered at this pixel is defined as the sum of the old gray-values of all pixels. Filtering the dilation can also be iterated. Multiple dilation steps with a given structuring element are equivalent to a single dilation step with a larger structuring element (only an up scaled version of the small structuring element for convex structuring elements). Similarly, erosion extends dark structures in the image by replacing the gray-value of each pixel with the minimum of old gray-values of all pixels within the structuring element [44,45,46,47].

Erosion: The erosion transformation of X by B is defined as the set of points x such that the translated Bx is contained in X and is expressed as below (1)

$$X\theta B=\left\{X|{B}_{X}\subseteq X\right\}$$

Erosion operation shrinks the image object. The degree of required shrinking and the direction of the shrinking can be controlled by defining the characteristics of the structuring element. If we reduce the size of the structuring element, the “harshness” of the erosion process reduces. Alternatively, various degrees of shrinking can be achieved by performing iterative erosions by considering the primitive element [48,49,50].

Dilation: The dilation transformation of X by B is defined as the set of points x such that the intersection of the translated structuring element Bx and image X is not a null set [51, 52]

$$X\oplus B=\left\{X|{B}_{X}\cap X\ne \varnothing \right\}$$

Opening: The multiscale opening of X by a structuring element B of size n is the combination of erosion followed by dilation by the structuring element nB [53].

$$XonB=\left(X\theta nB\right)\oplus nB,n=\mathrm{1,2},..,N.$$

Closing: Closing of X by a structuring element B is dilation by nB followed by erosion by nB. Opened image is the subset of main image but main image is the subset of closed image [54, 55].

$$X\cdot nB=\left(X\oplus nB\right)\theta nB,n=\mathrm{1,2},..,N.$$

Morphological operations—erosion, dilation, opening, and closing—serve as fundamental tools in processing stone inscription images. Erosion shrinks image structures, controlled by the structuring element, while dilation extends and highlights structures. Both erosion and dilation can be iteratively adjusted for fine-tuning. Opening, combining erosion and dilation, removes noise while preserving essential features, whereas closing fills gaps while maintaining overall integrity. These operations are pivotal in manipulating and refining image structures, enhancing the interpretation and analysis of crucial details within stone inscriptions.

Edge detection-laplacian, sobex, sobely and canny

Canny edge detection

It uses the first derivative of Gaussian to detect the edges of an images [56, 57]. The approach is based on convolution of the image function(f(x,y)) with the following Gaussian operator:

$$G\left(x,y;\sigma \right)=\frac{1}{2\pi {\sigma }^{2}}{e}^{-\left({x}^{2}+{y}^{2}\right)/2{\sigma }^{2}}$$

where \(\sigma\) is the spread of the Gaussian which controls the degree of smoothing. A new function \({f}{\prime}\left(x,y\right)\) is computed as

$${f}{\prime}\left(x,y\right)=G\left(x,y;\sigma \right)\times f(x,y)$$

Then, using the gradient of a pixel(x,y) in the \({f}{\prime}\left(x,y\right)\), the edges of the f(x,y) image can be detected.

Laplacian

LoG edge detector is based on the second order derivative of a Gaussian function. Consider the Gaussian function [58, 59].

$$G\left(x,y\right)={e}^{-\frac{{(x}^{2}{+y}^{2)}}{{2\partial }^{2}}}$$

The Laplacian of this function is

$${\nabla }^{2}G\left(x,y\right)=\frac{{\partial }^{2}G\left(x,y\right)}{{\partial x}^{2}}+\frac{{\partial }^{2}G\left(x,y\right)}{{\partial y}^{2}}=\left[\frac{{x}^{2}{+y}^{2-}{2\partial }^{2}}{{\partial }^{4}}\right]{e}^{-\frac{{(x}^{2}{+y}^{2)}}{{2\partial }^{2}}}$$

This function is called Laplacian of Gaussian (LoG). So, the image is convolved with this function and produces two effects- smoothing, and computing the Laplacian which yields a double-edge image. The edges are then detected by finding the zero crossings between the double edges.

Sobel

Functional derivative reflects the marked extent of image gradation variety. The Local Maximum of first derivative reflect the max extent of image gradation variety. The derivative value can be used as edge intensity value, so the edges can be detected by setting threshold [59].

$$G\left(x,y\right)=\left[\begin{array}{c}{G}_{x}\\ {G}_{y}\end{array}\right]=\left[\begin{array}{c}\frac{\partial f}{\partial x}\\ \frac{\partial f}{\partial y}\end{array}\right]$$

Sobel [60] operator utilizes two convolution kernels to calculate first-order derivative. The convolution kernels are

$${G}_{x}=\left[\begin{array}{ccc}-1& 0& 1\\ -2& 0& 2\\ -1& 0& 1\end{array}\right]{G}_{y}=\left[\begin{array}{ccc}1& 2& 1\\ 0& 0& 0\\ -1& -2& -1\end{array}\right]$$

In stone inscription analysis, these methods are applied to detect and highlight edges crucial for interpreting inscribed content. Canny Edge Detection's ability to preserve edges while smoothing the image aids in capturing intricate details. LoG's double-edge image and zero-crossing detection enhance edge detection accuracy. Sobel's use of derivative values helps discern edge intensities, facilitating edge detection by thresholding. Together, these methods contribute significantly to unveiling and analyzing crucial details within stone inscriptions.

K-means

Clustering or data grouping is a key initial procedure in image processing. K-means is typically used to locate objects and boundaries in images. It is used to find natural clusters within given data based upon varying input parameters. Clusters can be formed for images based on pixel intensity, color, texture, location, or some combination of these. The membership for each data point belongs to its nearest center, depending on the minimum distance [61].

$${c}_{j}=\frac{{\sum }_{i=1}^{n}m({c}_{j}\left|{x}_{i}\right)w\left({x}_{i}\right){x}_{i}}{{\sum }_{i=1}^{n}m({c}_{j}\left|{x}_{i}\right)w\left({x}_{i}\right)}$$
$$A\left(X,C\right)={\sum }_{i=1}^{n}{min}_{j=(1..i)}{{|x}_{i-}{c}_{j}|}^{2}$$

where centers is C and for each data point xi, compute its minimum distance with each center cj. For each center cj, recomputed the new center from all data points xi belong to this cluster.

In stone inscription analysis, K-means clustering can be applied to segment the inscribed content from the background or distinguish various elements within the images. It helps in identifying patterns, delineating different components, or separating text from non-text regions. By grouping pixels based on their characteristics, this technique facilitates subsequent analysis, interpretation, or feature extraction from stone inscription images.

Fast means denoising

Fast means denoising is replacing the color of a pixel with an average of the colors of similar pixels. NLM, neighborhood weightages are computed using the window similarity technique [62]. In this filter method, each pixel I′(\({x}_{i},{y}_{i}\)) is estimated as weighted mean of all the pixels in the image as shown in below equation

$${I}{\prime}\left({x}_{i},{y}_{i}\right)=\sum_{\left({x}_{i},{y}_{i}\right)\in \Omega }w\left(i,j\right)I\left({x}_{j},{y}_{j}\right)$$

where the weight ω (i, j) between two pixels (xi, yi) and (xj, yj) depends on their similarity as shown in below equation

$$w\left(i,j\right)=\frac{1}{Z\left(i\right)}\sum_{lvl\varepsilon levels}\begin{array}{c}{e}^{-{g}_{lvl\left[\frac{\left({I}_{lvl}\left({x}_{i},{y}_{i}\right)-{I}_{lvl}\left({x}_{j},{y}_{j}\right)\right){}^{2}}{{\alpha }_{lvl}{\sigma }^{2}}\right]}}\end{array}$$

where \(\left({I}_{lvl}\left({x}_{i},{y}_{i}\right)\right)\) is the pixel mean value of window of level lvl size which is centered at\(({x}_{i},{y}_{i}\)). σ is the standard deviation of the Gaussian noise of the input image, Ζ(i) is a normalized constant, \({g}_{lvl}\) is a gaussian weightage for image level lvl and \({\alpha }_{lvl}\) value is a scale factor to the computed image noise, which maps the noise variance to the corresponding image level lvl.

In stone inscription analysis, these denoising techniques—Fast Means and NLM—are instrumental in reducing noise, enhancing image clarity, and preserving critical details within the inscriptions. By averaging similar pixels or considering their weighted similarity, these methods effectively mitigate noise, allowing for clearer interpretation and analysis of the inscribed content. Table 1 displays a summarized comparison of various pre-processing techniques.

Table 1 Tabulated summary of all pre-processing technique

MSE

The mean square error (MSE) [62, 63] is the cumulative square error between the restored image and the original image defined as:

$$MSE=\frac{1}{MN}\sum_{y=1}^{M}\sum_{x=1}^{N}{\left[I\left(x,y\right)-{I}{\prime}\left(x,y\right)\right]}^{2}$$

where, M × N is the Image size, I(x,y) is an original image and I'(x,y) the restored image.

The peak signal-to-noise ratio (PSNR) is the peak value of signal-to-noise ratio (SNR) and in another words it is defined as the ratio of the maximum possible power of a pixel value and the power of distorting noise. As it known that, it affected the original image quality. It is defined as:

$$PSNR = 20 \times log_{{10}} \left( {\frac{{255}}{{\sqrt {MSE} }}} \right)$$

where, 255 × 255 is the maximum value of pixel present in an image and MSE is calculated for original and restored image with M × N size.

Results and discussion

In this experimental study, the dataset consists of images captured using a camera, featuring stone inscriptions from the Tanjore Brihadeeswar Temple dating back to the eleventh century, during the reign of Raja Raja Chola. A total of 200 stone inscription images were included in the dataset, having been previously scanned. These images are of a resolution of 200 × 200 dots per inch (dpi) and are in RGB format. Several sample images from the original dataset are illustrated in Fig. 3. The study encompasses various categories of images, such as Dark-illegible, Flawless-bright-illegible, Flawless-bright-legible, Flawless-irregular-moderate, highly impaired-irregular-illegible, Flawless-bright-legible, Flawless-bright-moderately legible, Flawless-dull-moderate, highly impaired-dark-illegible, Impaired-dark-moderate, Very impaired-dusky dark-legible, Impaired-dull-moderately legible, Impaired-dusky dark-moderate. To enhance the images for subsequent analysis, a series of preprocessing methods were employed, including Brightness and Contrast adjustment, Image Smoothening, Noise Removal, Structural Elements extraction, Separation of Dark and Light regions, Edge detection, and Fast denoising techniques. Separating the dark and light regions within the stone inscriptions is a critical task, and to achieve this, a range of methods were utilized, including Thresholding, Adaptive Thresholding using Gaussian, OSTU, Threshold Triangle, Adaptive Thresholding using Mean, Adaptive Thresholding using Gaussian (binary inversion), Adaptive Thresholding using Mean (binary inversion), Niblack, Savoula, Niblack Dilation, Niblack Morphology, Niblack Erosion, Savoula Dilation, Savoula Morphology, Savoula Erosion, Nilblack k-means, Nilblack k-means noising, Nilblack k-means noising (binarizations), and Nilblack l-means noising (binarization normalization). For the purpose of edge detection, Sobelx, SobelY, Canny edge detection, and Laplacian methods were employed. The Peak Signal-to-Noise Ratio (PSNR) values were computed for all image types following preprocessing, and these results are presented in Table 2 (Fig. 4).

Table 2 Comparative PSNR (Peak Signal-to-Noise Ratio) performance among various pre-processing methods
Fig. 4
figure 4

Sample Original Image and Binarized image MSE and PSNR result of diverse type of stone background structure inscriptions (Test case: 1. Flawless-bright-legible-1; 2. Flawless-bright-legible-2; 3. Flawless moderately-legible; 4. Flawless bright-moderately-legible; 5. Highly impaired dark irregular illegible; 6. Impaired-dull-moderately-legible; 7. Flawless dull moderate; 8. Dark legible)

The original input images were first transformed into grayscale, after which noise reduction was executed using the Gaussian filtering technique. Subsequently, various binarization techniques were applied. An example of the binarized outcomes obtained from the analysis of ancient stone inscriptions using the diverse preprocessing methods is visualized in Fig. 5. To ascertain the suitability of binarization techniques, a tenfold cross-validation process was executed for the 200 images extracted from the ancient stone inscriptions. The implementation of this experiment utilized OpenCV. Before script recognition, it's crucial to undertake image enhancement and restoration steps. Consequently, this experiment entailed the selection of 20 preprocessing filters. To gauge the image quality following preprocessing, metrics such as PSNR and Mean Squared Error (MSE) were computed. Through human visual evaluation, it was observed that neither Niblack nor Suavola algorithms stood out as superior. Therefore, a total of 34 algorithms were selected. The performance of the selection system is outlined in Table 2. The results of this experiment highlight that Adaptive Threshold Gaussian, Otsu's algorithm, and Nilblack k-means exhibited improved performance. When comparing outcomes derived from global thresholding methods with those from local adaptive thresholding methods, it was evident that local adaptive thresholding methods yielded a more stable character appearance by fine-tuning outputs at local levels, as opposed to global techniques [64,65,66]. While global thresholding methods managed to eliminate noise from certain background regions, they often rendered characters in other areas illegible [67]. The study involved the generation and preprocessing of stone inscription images, encompassing smoothening, noise reduction, filtering, structural element extraction, brightness enhancement, contrast enhancement, denoising, edge detection, and dark–light region separation. The investigation aimed to determine the most effective preprocessing technique by calculating PSNR and MSE values. It was noted that for images with a dark background, the median filter and the Niblack method outperformed other preprocessing techniques. Recent research has highlighted Adaptive Gaussian thresholding as effective for sharp bright images [40], which aligns with the findings of this study for primitive pillar stone inscriptions that are typically dull. Ancient Tamil stone inscriptions have proven challenging to interpret due to their high number of characters with subtle differences [68]. Research also suggests that performance metrics like PSNR play a pivotal role in assessing existing preprocessing methods [69], with denoising interventions having a positive impact on image resolution and character recognition [70].

Fig. 5
figure 5

Results achieved through binarization with the utilization of various preprocessing approaches. A Original image, B Brightness and contrast, C Gray, D Blur, E Gaussian blur, F Median blur, G Bilateral, H Adaptive Gaussian, I Binary, J OSTU, K Binary triangle, L Adaptive mean, M Adaptive gaussian bin inv, N Adaptive mean bin inv, O Niblack, P Sauvoula, Q Adaptive Gaussian morpological line, R Erosion, S dilation, T Adaptive Gaussian k-means, U Adaptive Gaussian k-means fast denoising, V Binarization, W Normalize, X Canny edge, Y Canny edge Thinning, Z Sobel x, AA Sobel y, BB Lapclacian, CC Equalizer histogram, DD Equalizer histogram threshold

It sounds like the preprocessing steps are critical in achieving clear and complete character extraction from the background. From what you've described, the order of operations for preprocessing is crucial in determining the success of separating foreground characters from the background. It appears that the combination of preprocessing steps involving brightness and contrast adjustment, grayscale conversion, resizing, median and Gaussian blur, erosion and dilation, adaptive thresholding (like Gaussian and Sauvola), followed by k-means denoising, leads to successful character extraction without breakage or discontinuity in the flawless-bright-legible/moderate background scripts shown in Fig. 4a–d. However, when the same preprocessing steps are applied to darker or more irregular backgrounds (such as dark/dusky dark –legible/moderate and impaired /illegible /dull/irregular backgrounds), the resulting images show added noise, broken character pixels, discontinuity, and meshed or meaningless character retrieval as shown in Fig. 4e–h. This poor result leads to a lower recognition rate. It seems that the challenge lies in adapting the preprocessing techniques to different types of backgrounds, especially when dealing with darker, irregular, or noisy backgrounds. Further optimization or adjustment of the preprocessing steps might be necessary to handle such variations in backgrounds for better character extraction and recognition, especially in cases with more challenging backgrounds.

Limitations of current work: The study’s primary limitation revolves around its focused exploration of preprocessing methods. While it examines numerous techniques, it doesn’t encompass the entire array of available algorithms and methodologies. This selective approach might have overlooked potentially effective techniques or synergistic combinations that could offer improved outcomes but were not included in this study. Additionally, while the study successfully applies various preprocessing methods to enhance image quality and facilitate character recognition in ancient Tamil inscriptions, it primarily focuses on the preprocessing aspect rather than directly addressing script recognition. This implies that while the images are improved for analysis, the study does not delve deeply into the accuracy or efficacy of recognizing and interpreting characters within these inscriptions. Moreover, the evaluation metrics, like PSNR and MSE, while informative about image quality improvement, might not completely capture the nuances of text legibility or character recognition accuracy, which are critical for historical script analysis. Lastly, the study doesn’t extensively address the optimization of parameters within each preprocessing method. Optimizing these parameters could potentially yield better results in terms of character recognition or text legibility for ancient scripts.

In summary, while the research makes significant strides in enhancing stone inscription images, it is constrained by the limited scope of preprocessing methods assessed and the lack of direct analysis on character recognition accuracy. This leaves room for future investigations to explore a wider array of preprocessing techniques and to deeply assess the effectiveness of these methods in improving the interpretation of ancient scripts.

Conclusion

Overall 34 pre-processing methods were assessed for best PSNR values for different type of images (Dark-illegible, Flawless-bright-illegible, Flawless-bright-legible, Flawless-irregular-moderate, highly impaired-irregular-illegible, Flawless-bright-legible, Flawless-bright-moderately legible, Flawless-dull-moderate, Highly impaired-dark-illegible, Impaired-dark-moderate, Very impaired-dusky dark-legible, Impaired-dull-moderately legible, Impaired-dusky dark-moderate). In the realm of Tamil literature, the task of identifying characters is of paramount importance, although it remains a challenging and intricate endeavor due to the presence of numerous characters in ancient Tamil scripts that exhibit resemblances or subtle deviations. Within the scope of this investigation, it was revealed that employing a combined approach involving the Nilblack-k-means method effectively handles the inscription processing, ultimately yielding processed images of elevated resolution.

Data availability

The datasets used and/or analysed during the current study available from the corresponding author on reasonable request.

References

  1. Devi KD, Maheswari PU. Insight on character recognition for calligraphy digitization. In 2017 IEEE Technological Innovations in ICT for Agriculture and Rural Development (TIAR). IEEE. 2017. pp. 78–83.

  2. Bhuvaneswari G, Subbiah Bharathi V. An efficient positional algorithm for recognition of ancient stone inscription characters. In: Advanced Computing (ICoAC) 2015 Seventh International Conference on. IEEE. 2015.

  3. Janani G, Vishalini V, Mohan Kumar P. Recognition and analysis of Tamil inscriptions and mapping using image processing techniques. Science Technology Engineering and Management (ICONSTEM) Second International Conference on. IEEE. 2016.

  4. Devi K, Maheswari PU. Digital acquisition and character extraction from stone inscription images using modified fuzzy entropy-based adaptive thresholding. Soft Comput. 2019;23:2611–26.

    Article  Google Scholar 

  5. Mahalakshmi M, Sharavanan M. Ancient Tamil script recognition and translation using LabVIEW. In International conference on communication and signal processing. 2013. pp. 1021–6.

  6. Vellingiriraj EK, Balamurugan M, Balasubramanie P. Text analysis and information retrieval of historical Tamil ancient documents using machine translation in image zoning. Int J Lang Lit Linguist. 2016;2(4):164–8.

    Google Scholar 

  7. Chaki N, Shaikh SH, Saeed K. A comprehensive survey on image binarization techniques. In: Chaki N, editor. Exploring image binarization techniques. Berlin: Springer; 2014. p. 5–15.

    Chapter  Google Scholar 

  8. Pal U, Roy PP, Tripathy N, Lladós J. Multi-oriented Bangla and Devanagari text recognition. Pattern Recogn. 2010;43(12):4124–36.

    Article  ADS  Google Scholar 

  9. Durga Devi K, Maheswari PU, Polasi PK, Preetha R, Vidhyalakshmi M. Pattern matching model for recognition of stone inscription characters. Comput J. 2023;66(3):554–64.

    Article  MathSciNet  Google Scholar 

  10. Buzykanov SN. Enhancement of poor resolution text images in the weighted Sobolev space. In 2012 19th International Conference on Systems, Signals and Image Processing (IWSSIP). IEEE. 2012. pp. 536–9.

  11. Kavallieratou E, Antonopoulou H. Cleaning and enhancing historical document images. Lect Notes Comput Sci. 2005;3708:681–8.

    Article  Google Scholar 

  12. Seeger M, Dance C. Binarising camera images for OCR. In: ICDAR 2001. 2001. pp. 54–9.

  13. Pal U, Chaudhuri BB. Indian script character recognition: a survey. Pattern Recogn. 2004;37:1887–99.

    Article  Google Scholar 

  14. Vasantha Lakshmi C, Patvardhan C. An optical character recognition system for printed Telugu text. Pattern Anal Appl. 2004;7(2):190–204.

    MathSciNet  Google Scholar 

  15. Ptak R, Żygadło B, Unold O. Projection-based text line segmentation with a variable threshold. Int J Appl Math Comput Sci. 2017;27:101–414.

    Article  MathSciNet  Google Scholar 

  16. Panyam NS, Vijaya Lakshmi TR, Krishnan R, Koteswara Rao NV. Modeling of palm leaf character recognition system using transform based techniques. Pattern Recogn Lett. 2016;84:29–34.

    Article  ADS  Google Scholar 

  17. Goh TY, Basah SN, Yazid H, Safar MJA, Ahmad Saad FS. Performance analysis of image thresholding: Otsu technique. Measurement. 2018;114:298–307.

    Article  ADS  Google Scholar 

  18. Ripon S, Chowdhury L, Ashour AS, Dey N. Machine-learning approach for ribonucleic acid primary and secondary structure prediction from images. In: Dey N, Ashour AS, Shi F, Balas VE, editors. Soft computing based medical image analysis. Cambridge: Academic Press; 2018. p. 203–21.

    Chapter  Google Scholar 

  19. Davies ER. The role of thresholding. In: Davies ER, editor. Computer vision. 5th ed. Cambridge: Academic Press; 2018. p. 93–118.

    Chapter  Google Scholar 

  20. Siddique MAB, Arif RB, Khan MMR. Digital Image Segmentation in Matlab: a Brief Study on OTSU’s Image Thresholding. In 2018 International Conference on Innovation in Engineering and Technology (ICIET). IEEE. 2018. pp. 1–5.

  21. Otsu N. A threshold selection method from gray-level histograms. IEEE Trans Syst Man Cybern. 1979;9(1):62–6.

    Article  Google Scholar 

  22. Su B, Lu S, Tan CL. Binarization of historical document images using the local maximum and minimum. In: Proc Intl Workshop on Document Analysis Systems, 2010. pp. 159–65.

  23. Khurshid K, Siddiqi I. Comparison of Niblack inspired Binarization methods for ancient documents. In: Proceedings of SPIE, 2009. pp. 1–10.

  24. Saxena LP. Niblack’s binarization method and its modifications for real-time applications: a review. Artif Intell Rev. 2017;47(4):469–98.

    Google Scholar 

  25. Sudarsan D, Sankar D. A Novel complete denoising solution for old Malayalam palm leaf manuscripts. Pattern Recognit Image Anal. 2022;32(1):187–204.

    Article  Google Scholar 

  26. Sezgin M, Sankur B. Survey over image thresholding techniques and quantitative performance evaluation. J Electron Imaging. 2004;13(1):146–65.

    Article  ADS  Google Scholar 

  27. Bovik AC. Streaking in median filtered images. IEEE Trans Acoust Speech Signal Process. 1987;35(4):493–503.

    Article  ADS  Google Scholar 

  28. Stark JA. Adaptive image contrast enhancement using generalizations of histogram equalization. IEEE Trans Image Process. 2000;9(5):889–96.

    Article  ADS  CAS  PubMed  Google Scholar 

  29. Yadav G, Maheshwari S, Agarwal A. Contrast limited adaptive histogram equalization based enhancement for real-time video system. In: 2014 International Conference on Advances in Computing, Communications and Informatics (ICACCI), 2014. 2392–7.

  30. Bedi S, Khandelwal R. Various image enhancement techniques: a critical review. Int J Adv Res Computer Commun Eng. 2013;2(3):1605–9.

    Google Scholar 

  31. Cheng H, Shi X, Tan CL. A simple and effective histogram equalization approach to image enhancement. IEEE Trans Image Process. 2004;14(2):158–70.

    Google Scholar 

  32. Huang SC, Cheng FC, Chiu YS. Efficient contrast enhancement using adaptive gamma correction with weighting distribution. IEEE Trans Image Process. 2013;22(3):1032–41.

    Article  ADS  MathSciNet  PubMed  Google Scholar 

  33. Chaki N. Exploring image Binarization techniques. Berlin: Springer; 2014.

    Book  Google Scholar 

  34. Park GH, Cho HH, Choi MR. A contrast enhancement method using dynamic range separate histogram equalization. IEEE Trans Consum Electron. 2008;54(4):2067–74.

    Article  Google Scholar 

  35. Goh TY, Basah SN, Xue X. Fog removal from video sequences using contrast limited adaptive histogram equalization. Computational Intelligence and Software Engineering 2009. CiSE 2009. International Conference. 2009. pp. 1–4.

  36. Jin Y, Laura M, Laine Fayad A. Contrast enhancement by multi scale adaptive histogram equalization. Proc SPIE. 2001;4478:206–13.

    Article  ADS  Google Scholar 

  37. Ntirogiannis K, Gatos B, Pratikakis I. Performance evaluation methodology for historical document image binarization. IEEE Trans Image Process. 2013;22(2):595–609.

    Article  ADS  MathSciNet  PubMed  Google Scholar 

  38. Pratikakis I, Zagoris K, Kaddas P, Gatos B. ICFHR 2018 competition on handwritten document image binarization (H-DIBCO 2018). In: 2018 16th International Conference on Frontiers in Handwriting Recognition (ICFHR). 2018. pp. 489–93.

  39. Zimmerman JB, Pizer SM, Staab EV, Perry JR, McCartney W, Brenton BC. An evaluation of the effectiveness of adaptive histogram equalization for contrast enhancement. IEEE Trans Med Imaging. 1988;7(4):304–12.

    Article  CAS  PubMed  Google Scholar 

  40. Rahman NA, Haroon F. Adaptive Gaussian and double thresholding for contour detection and character recognition of two-dimensional area using computer vision. Eng Proc. 2023;32(1):23.

    Google Scholar 

  41. Durrand F, Dorsey J. Fast bilateral filtering for the display of high dynamic range images. In: Proceedings of SIGGRAPH 2002. 2002. pp. 844–7.

  42. Durand F, Dorsey J. Fast bilateral filtering for the display of high dynamic-range images. ACM Trans Graph. 2002;21(3):257–66.

    Article  Google Scholar 

  43. Paris S, Durand F. A fast approximation of the bilateral filter using a signal processing approach. Int J Comput Vision. 2009;81:24–52.

    Article  Google Scholar 

  44. Tcheslavski GV. Morphological image processing: grayscale morphology. ELEN 4304/5365 DIP, Spring 2010. 2010.

  45. Déforges O, Normand N, Babel M. Fast recursive grayscale morphology operators: from the algorithm to the pipeline architecture. J Real-Time Image Proc. 2013;8(2):143–52.

    Article  Google Scholar 

  46. Srisha R, Khan A. Morphological operations for image processing: understanding and its applications. 2013.

  47. Clienti C, Beucher S, Bilodeau M. A system on chip dedicated to pipeline neighborhood processing for mathematical morphology. In: IEEE Conference in Signal Processing, 16th European, 1–5. 2008.

  48. Torres-Huitzil C. Fast hardware architecture for grey level image morphology with flat structuring elements. IET Image Proc. 2013;8(2):112–21.

    Article  Google Scholar 

  49. Heijmans H. Morphological image operators. In: Marton L, editor. Advances in electronics and electron physics. Cambridge: Academic Press; 1994.

    Google Scholar 

  50. Haralick R, Sternberg S, Zhuang X. Image analysis using mathematical morphology. IEEE Trans Pattern Anal Mach Intell. 1987;9(4):532–50.

    Article  CAS  PubMed  Google Scholar 

  51. Bartovský J, Dokládal P, Dokládalová E, Georgiev V. Parallel implementation of sequential morphological filters. J Real-Time Image Proc. 2014;9(2):315–27.

    Article  Google Scholar 

  52. Gil J, Kimmel R. Efficient dilation, erosion, opening, and closing algorithms. IEEE Trans Pattern Anal Mach Intell. 2002;24(12):1606–17.

    Article  Google Scholar 

  53. Gil J, Kimmel R. Efficient dilation, erosion, opening and closing algorithms in mathematical morphology and its applications to image and signal processing. In: Goutsias J, Vincent L, Bloomberg D, editors. Proceedings of Shape in Picture ‘92, NATO Workshop, Driebergen, The Netherlands, September 1992. Springer-Verlag. 2000. pp. 301–10.

  54. Vincent L. Morphological area openings and closings for greyscale images. In: Proceedings of Shape in Picture ‘92, NATO Workshop, Driebergen, The Netherlands. 1992.

  55. Dokládal P, Dokladalova E. Computationally efficient, one-pass algorithm for morphological filters. J Vis Commun Image Represent. 2011;22(5):411–20.

    Article  Google Scholar 

  56. Gonzalez CI, Melin P, Castro JR, Castillo O. Edge detection methods and filters used on digital image processing. In: Gonzalez CI, Melin P, Castro JR, Castillo O, editors. Edge detection methods based on generalized type-2 fuzzy logic. Berlin: Springer; 2017.

    Chapter  Google Scholar 

  57. Mutneja V. Methods of image edge detection: a review. J Electr Electron Syst. 2015;4:2332–796.

    Google Scholar 

  58. Gentsos C, Sotiropoulou C, Nikolaidis S, Vassiliadis N. Realtime canny edge detection parallel implementation for FPGAs. In: Proceedings of the International Conference on Electronics, Circuits and Systems. 2010. pp. 499–502.

  59. Chao L, Jiliu Z, Kun H. Adaptive edge-detection method based on canny algorithm. Comput Eng Design. 2010;31(18):4036–9.

    Google Scholar 

  60. Vincent OR. A descriptive algorithm for Sobel image edge detection. In: Proceedings of Informing Science & IT Education Conference (InSITE). 2009.

  61. Jayanthi N, Sharma T, Sharma V, Tyagi S, Indu S. Classification of ancient inscription images on the basis of material of the inscriptions. In: 2021 3rd International Conference on Signal Processing and Communication (ICPSC), 2021. pp. 422–7.

  62. Vijayalakshmi R, Gnanasekar JM. A review on character recognition and information retrieval from ancient inscriptions. In: 2022 8th International Conference on Smart Structures and Systems (ICSSS), 2022. pp. 1–7.

  63. Dhivya S, Beulah JR. Ancient Tamil character recognition from stone inscriptions—a theoretical analysis. In: 2022 2nd Asian Conference on Innovation in Technology (ASIANCON), 2022. pp. 1–8.

  64. Rajithkumar BK, Mohana HS, Uday J, Bhavana MB, Anusha LS. Read and recognition of old Kannada stone inscriptions characters using a novel algorithm. In: 2015 International Conference on Control, Instrumentation, Communication and Computational Technologies (ICCICCT), 2015. pp. 284–88.

  65. RajaKumar S, Subbiah Bharathi V. Eighth century Tamil consonants recognition from stone inscriptions. Int Conf Recent Trends Inform Technol. 2012;2012:40–3.

    Google Scholar 

  66. Rajnish P, Kamath KP, Kumar B, Nishanth M, Preethi P. Improving the quality and readability of ancient Brahmi stone inscriptions. In: 2023 2nd International Conference for Innovation in Technology (INOCON), 2023. pp. 1–8.

  67. Rogowska J. Overview and fundamentals of medical image segmentation. In: Bankman IN, editor. Handbook of medical image processing and analysis. 2nd ed. Cambridge: Academic Press; 2009. p. 73–90.

    Google Scholar 

  68. Priya RD, Karthikeyan S, Indra J, Kirubashankar S, Abraham A, Gabralla LA, Nandhagopal SM. Self-adaptive hybridized lion optimization algorithm with transfer learning for ancient Tamil character recognition in stone inscriptions. IEEE Access. 2023. https://doi.org/10.1109/ACCESS.2023.3268545.

    Article  Google Scholar 

  69. Sukanthi S, Murugan SS, Hanis S. Binarization of stone inscription images by modified bi-level entropy thresholding. Fluct Noise Lett. 2021;20(06):2150054.

    Article  ADS  Google Scholar 

  70. Zhang H, Qi Y, Xue X, Nan Y. Ancient stone inscription image denoising and inpainting methods based on deep neural networks. Discret Dyn Nat Soc. 2021;2021:1–11.

    Google Scholar 

Download references

Acknowledgements

The first author is thankful to Anna University Chennai for supporting the research.

Funding

Not applicable.

Author information

Authors and Affiliations

Authors

Contributions

Conceptualization, JJ and PUM. Methodology, JJ and PUM. Investigation, JJ; writing original draft preparation, JJ; writing review and editing JJ and PUM.

Corresponding author

Correspondence to J. Jayanthi.

Ethics declarations

Ethical approval and consent to participate

Not applicable.

Competing interests

The authors declare no competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Jayanthi, J., Maheswari, P.U. Comparative study: enhancing legibility of ancient Indian script images from diverse stone background structures using 34 different pre-processing methods. Herit Sci 12, 63 (2024). https://doi.org/10.1186/s40494-024-01169-6

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s40494-024-01169-6

Keywords