Notice Board :

Call for Paper
Vol. 4 Issue 5

Submission Start Date:
May 01, 2018

Acceptence Notification Start:
May 10, 2018

Submission End:
May 15, 2018

Final MenuScript Due:
May 25, 2018

Publication Date:
May 31, 2018


                         Notice Board: Call for PaperVol. 4 Issue 5      Submission Start Date: May 01, 2018      Acceptence Notification Start: May 10, 2018      Submission End: May 15, 2018      Final MenuScript Due: May 25, 2018      Publication Date: May 31, 2018




Volume II Issue I

Author Name
Neha S Patil, Prof Chhaya Nayak
Year Of Publication
2016
Volume and Issue
Volume 2 Issue 1
Abstract
To locate the suitable number of bunches and to apportion the archives is urgent in report grouping. In this paper we will concentrate on different bunching strategies and our proposed framework is to find the group structure without giving the aggregate number of groups as information. Report elements or even we can say that the different characteristics will be with no human obstruction isolated into two gatherings, specifically, discriminative words and nondiscriminative words, and contribute diversely to record grouping. There is variational surmising calculation in which we derive the archive accumulation structure and words in the meantime parcel of report. our proposed approach for the semisupervised report bunching. Semi-administered grouping lies between both programmed order and auto-association. Here the manager need not indicates an arrangement of classes, but rather just to give an arrangement of writings gathered by the criteria to be utilized to produce Clusters.
PaperID
2016/IJRRETAS/2/2016/1611

Author Name
Rishikesh S Patil, Prof Chhaya Nayak
Year Of Publication
2016
Volume and Issue
Volume 2 Issue 1
Abstract
Video is becoming a prevalent medium for e-learning. Lecture videos contain text information in both the visual and aural channels: the presentation slides and lecturer's speech. This paper examines the relative utility of automatically recovered text from these sources for lecture video retrieval. To extract the visual information, we apply video content analysis to detect slides and optical character recognition to obtain their text. We extract textual metadata by applying video Optical Character Recognition (OCR) technology on key-frames and Automatic Speech Recognition (ASR) on lecture audio tracks. The OCR and ASR transcript as well as detected slide text line types are adopted for keyword extraction, by which both video- and segment-level keywords are extracted for content-based video browsing and search.
PaperID
2016/IJRRETAS/2/2016/1612