Work place: BBD University Lucknow, India
E-mail: avinash.verma93@yahoo.com
Website:
Research Interests: Image Processing, Image Manipulation, Image Compression, Computer systems and computational processes
Biography
Avinash Verma born in India 1988 has Done B.tech in Computer Science and Engineering, M.Tech in Information Technology from CDAC Noida. Qualifying GATE 2010 and 2016 and currently perusing PhD in Computer Science and Engineering. His area of interest is Image processing and Optical character Recognitions. He has published papers in National and Internationals Journals.
By Avinash Verma Deepak Kumar Singh
DOI: https://doi.org/10.5815/ijigsp.2017.10.04, Pub. Date: 8 Oct. 2017
The Main objective of this assistive framework is to communicate the textual Information in the image captured by the Visually Challenged person as Speech, So that the Visually Challenged person can acquire knowledge about the surrounding. This framework can help Visually Challenged person to read books, magazine, warnings, instructions and various displays as well by taking their image along with the surrounding. Then the Optical Character Recognition (OCR) extracts and recognizes the text in the image and generates the text file. This text file is further converted to Speech with the help of Text to Speech (TTS) Synthesis. The inherent problem with the previous approach was if the acquired image is affected with the issues of different lighting conditions, noise and issue of Skew and Blur, as the image is captured by Visually Challenged person. Then the overall accuracy of the system was at stake due to inefficient OCR leads to improper Speech output of TTS Synthesis. In this paper we have introduced two more processes that are deblurring using Blind Deconvolution method and Pre-processing operation to remove the effect of noise and blur. Thus it prepares the image for efficient result of the framework for Visually Challenged. The proposed approach is implemented in Matlab with the image captured manually and taken from the internet and the result along with the OCR text file and corresponding output Speech shows that our framework is better than the previous framework.
[...] Read more.By Avinash Verma Deepak Kumar Singh
DOI: https://doi.org/10.5815/ijigsp.2017.01.05, Pub. Date: 8 Jan. 2017
Objective of this paper is to propose a new Deblurring method for motion blurred textual images. This technique is based on estimating the blur kernel or the Point Spread Function of the motion blur using Blind Deconvolution method. Motion blur is either due to the movement of the camera or the object at the time of image capture. The point spread function of the motion blur is governed by two parameters length of the motion and the angle of the motion. In this approach we have estimated point spread function for the motion blur iteratively for different values of the length and angle of motion. For every estimated PSF we perform the Deconvolution operation with the blurred image to get the non- blurred or the latent image. Latent image obtained is then feed to an Optical character recognition so that the text in that image can be recognized. Then we calculate the Average Word Confidence for the recognized text. Thus for every estimated Point Spread Function and the obtained latent image we get the value of Average Word Confidence. The Point Spread Function with the highest Average Word Confidence value is the optimal Point Spread Function which can be used to deblur the given textual image. In this method we do not have any prior information about the PSF and only single image is used as an input to the system. This method has been tested with the naturally blurred image taken manually and through the internet as well as artificially blurred image for the evaluation of the results. The implementation of the proposed algorithm has been done in MATLAB.
[...] Read more.Subscribe to receive issue release notifications and newsletters from MECS Press journals