{"id":599,"date":"2016-08-09T11:54:18","date_gmt":"2016-08-09T11:54:18","guid":{"rendered":"http:\/\/doceng2016.caa.tuwien.ac.at\/?page_id=599"},"modified":"2016-08-24T14:25:57","modified_gmt":"2016-08-24T14:25:57","slug":"friday-september-16th","status":"publish","type":"page","link":"https:\/\/doceng2016.cvl.tuwien.ac.at\/?page_id=599","title":{"rendered":"Friday, September 16th"},"content":{"rendered":"<div id=\"DocEng16Program\">\n<table class=\"page_table\" cellpadding=\"0\" cellspacing=\"0\">\n<tbody>\n<tr>\n<td class=\"left_td_column\">       &nbsp; <\/p>\n<div class=\"left_spacer\">&nbsp;<\/div>\n<\/td>\n<td class=\"central_td_column\">\n<div style=\"visibility:hidden;margin:0pt;border:0pt;padding:0pt\" id=\"menu2\"> <\/div>\n<div id=\"menu3\" style=\"visibility:hidden;margin:0pt;border:0pt;padding:0pt\"> <\/div>\n<div id=\"content\">\n<p class=\"days\"> <a href=\"https:\/\/doceng2016.caa.tuwien.ac.at\/?page_id=583\" class=\"program_days\">Overview<\/a>| <a href=\"https:\/\/doceng2016.cvl.tuwien.ac.at\/?page_id=594\" class=\"program_days\">Wednesday<\/a>| <a class=\"program_days\" href=\"https:\/\/doceng2016.caa.tuwien.ac.at\/?page_id=598\">Thursday<\/a>| <a class=\"active_day program_days\" href=\"https:\/\/doceng2016.caa.tuwien.ac.at\/?page_id=599\">Friday<\/a> <\/p>\n<p>  <script src=\"styles\/program.js\"> <\/script> <script>addEventHandler(window,'load',function () {Program.ping1('\/statistics\/page_access_x.cgi',Program.ping2,'m1=1')})<\/script> <script>Program.data={pr:540,co:148955,pk:'pd:148955:2016-09-16'}<\/script> <\/p>\n<div class=\"session\"> <a name=\"session:10876\"> <\/p>\n<div class=\"heading\"> <span class=\"interval\">09:30-10:45<\/span> <span class=\"title\">Session 15: Applications &amp; Security<\/span> <\/div>\n<p> <\/a> <\/p>\n<div class=\"session_chair\">\n<div class=\"chair_word\">Chair: <\/div>\n<div class=\"chair_names\"> <a href=\"http:\/\/easychair.org\/smart-program\/DocEng2016\/person7.html\" class=\"person\">David F. Brailsford<\/a> <\/div>\n<\/p><\/div>\n<table class=\"talks\">\n<tbody>\n<tr class=\"talk\">\n<td class=\"time\">09:30<\/td>\n<td> <a name=\"talk:30524\"\/> <\/p>\n<div class=\"authors\"> <a name=\"talk:30524\"\/> <a href=\"http:\/\/easychair.org\/smart-program\/DocEng2016\/person71.html\" class=\"person\">Rodrigo Laiola Guimaraes<\/a>, <a class=\"person\" href=\"http:\/\/easychair.org\/smart-program\/DocEng2016\/person72.html\">Priscilla Avegliano<\/a> and <a href=\"http:\/\/easychair.org\/smart-program\/DocEng2016\/person73.html\" class=\"person\">Lucas Villa Real<\/a> <\/div>\n<div class=\"title\">A Lightweight and Efficient Mechanism for Fixing the Synchronization of Misaligned Subtitle Documents<\/div>\n<div class=\"speaker\"\/>\n<div class=\"abstract\">\n<p>ABSTRACT.  Online subtitle databases allow users to easily find subtitle documents  in multiple languages for thousands of films and TV series episodes. However, getting the subtitle document that gives satisfactory synchronization on the first attempt is like hitting the jackpot. The truth is that this process often involves a lot of trial-and-error because multiple versions of subtitle documents have distinct synchronization references, given that they are targeted at variations of the same audiovisual content. Building on our previous efforts to address this problem, in this paper we formalize and validate a two-phase subtitle synchronization framework. The benefit over current approaches lays in the usage of audio fingerprint annotations generated from the base audio signal as second-level synchronization anchors. This  way, we allow the media player to dynamically fix during playback the most common cases of subtitle synchronization misalignment that compromise users\u2019 watching experience. Results from our evaluation process indicate that our framework has minimal impact on existing subtitle documents and formats as well as on the playback performance.<\/p>\n<\/p><\/div>\n<\/td>\n<\/tr>\n<tr class=\"talk\">\n<td class=\"time\">10:00<\/td>\n<td> <a name=\"talk:30535\"\/> <\/p>\n<div class=\"authors\"> <a name=\"talk:30535\"\/> <a class=\"person\" href=\"http:\/\/easychair.org\/smart-program\/DocEng2016\/person197.html\">Laurent Denoue<\/a>, <a href=\"http:\/\/easychair.org\/smart-program\/DocEng2016\/person198.html\" class=\"person\">Scott Carter<\/a> and <a href=\"http:\/\/easychair.org\/smart-program\/DocEng2016\/person199.html\" class=\"person\">Matthew Cooper<\/a> <\/div>\n<div class=\"title\">DocuGram: Turning Screen Recordings into Documents<\/div>\n<div class=\"speaker\"\/>\n<div class=\"abstract\">\n<p>ABSTRACT.  In this paper we describe DocuGram, a novel tool to capture and share documents originating from any application. As users scroll through pages of their document inside the native application (Word, Google Docs, web browser), the system captures and analyses in real-time the rendered video frames and reconstitutes the original document pages into  an easy to view HTML-based representation. In addition to detecting and  regenerating the document pages, a DocuGram also includes the interactions users had over them, e.g. mouse motions and voice comments.  A DocuGram allows users to flexibly share enhanced documents across applications.<\/p>\n<\/p><\/div>\n<\/td>\n<\/tr>\n<tr class=\"talk\">\n<td class=\"time\">10:15<\/td>\n<td> <a name=\"talk:30538\"\/> <\/p>\n<div class=\"authors\"> <a name=\"talk:30538\"\/> <a class=\"person\" href=\"http:\/\/easychair.org\/smart-program\/DocEng2016\/person178.html\">Eya Ben Charrada<\/a> and <a class=\"person\" href=\"http:\/\/easychair.org\/smart-program\/DocEng2016\/person179.html\">Stefan Mussato<\/a> <\/div>\n<div class=\"title\">An Exploratory Study on Managing and Searching for Documents in Software Engineering Environments<\/div>\n<div class=\"speaker\"\/>\n<div class=\"abstract\">\n<p>ABSTRACT.  Large number of documents are usually produced in the software industry. In this work, we conduct a qualitative study to explore the main practices and challenges related to managing these documents. The results of this study are based on interviews with 13 practitioners from  nine companies. The main findings of the study are: (1) much data is stored in e-mails and in meeting protocols, (2) practitioners like wikis, (3) when searching for documents, practitioners would rather  browse the structure than use the search function and (4) searching for  documents is still a challenge due to the low effectiveness of search functions and the scattering of documents over several locations and tools.<\/p>\n<\/p><\/div>\n<\/td>\n<\/tr>\n<tr class=\"talk\">\n<td class=\"time\">10:30<\/td>\n<td> <a name=\"talk:30527\"\/> <\/p>\n<div class=\"authors\"> <a name=\"talk:30527\"\/> <a href=\"http:\/\/easychair.org\/smart-program\/DocEng2016\/person55.html\" class=\"person\">Margaret Sturgill<\/a> and <a class=\"person\" href=\"http:\/\/easychair.org\/smart-program\/DocEng2016\/person53.html\">Steven Simske<\/a> <\/div>\n<div class=\"title\">Mass Serialization Method for Document Encryption Policy Enforcement<\/div>\n<div class=\"speaker\"\/>\n<div class=\"abstract\">\n<p>ABSTRACT.  Analytics obtained during the creation of a database of mass serialized  codes can also be used to help enforcement of encryption policy on documents. In this paper, we introduce a set of metrics which complement  traditional NIST cryptography methods \u2013 4 mass serialization and one entropy metric &#8212; which in combination can allow a discrimination between encrypted vs. zipped files. We describe the use of these methods  to identify a broad range of non-randomness in number sets, and apply them to a more mundane problem\u2014that of automatic assessment of  the encryption state of a corpora of documents.<\/p>\n<\/p><\/div>\n<\/td>\n<\/tr>\n<\/tbody>\n<\/table><\/div>\n<div class=\"session\">\n<div class=\"coffeebreak\"> <span class=\"interval\">10:45-11:15<\/span> <span class=\"title\">Coffee Break<\/span> <\/div>\n<\/p><\/div>\n<div class=\"session\"> <a name=\"session:10870\"> <\/p>\n<div class=\"heading\"> <span class=\"interval\">11:15-12:45<\/span> <span class=\"title\">Session 16: Visual Document Analysis<\/span> <\/div>\n<p> <\/a> <\/p>\n<div class=\"session_chair\">\n<div class=\"chair_word\">Chair: <\/div>\n<div class=\"chair_names\"> <a href=\"http:\/\/easychair.org\/smart-program\/DocEng2016\/person53.html\" class=\"person\">Steven Simske<\/a> <\/div>\n<\/p><\/div>\n<table class=\"talks\">\n<tbody>\n<tr class=\"talk\">\n<td class=\"time\">11:15<\/td>\n<td> <a name=\"talk:30514\"\/> <\/p>\n<div class=\"authors\"> <a name=\"talk:30514\"\/> <a href=\"http:\/\/easychair.org\/smart-program\/DocEng2016\/person156.html\" class=\"person\">Prerana Jana<\/a>, <a href=\"http:\/\/easychair.org\/smart-program\/DocEng2016\/person157.html\" class=\"person\">Anubhab Majumdar<\/a>, <a href=\"http:\/\/easychair.org\/smart-program\/DocEng2016\/person158.html\" class=\"person\">Sekhar Mandal<\/a> and <a class=\"person\" href=\"http:\/\/easychair.org\/smart-program\/DocEng2016\/person159.html\">Bhabatosh Chanda<\/a> <\/div>\n<div class=\"title\">Generation of Search-able PDF of the Chemical Equations segmented from Document Images<\/div>\n<div class=\"speaker\"\/>\n<div class=\"abstract\">\n<p>ABSTRACT.  PDF format of scanned document images is not searchable. OCR tries to remedy this adversity by converting document images  into editable and searchable data, but it has it&#8217;s own limitations in presence of equations &#8211; both mathematical and chemical. OCR system for mathematical equation  is already a major research area and has provided  successful result. However, chemical equation segmentation has been a less ventured road. In this paper, we present a novel method for automated generation of searchable PDF format of segmented chemical equations from scanned document images by performing chemical symbol recognition and auto-correction of OCR output. We use existing OCR system, pattern recognition technique, contextual data analysis and a standard LaTeX package to generate the chemical equation in searchable PDF format. The effectiveness of the proposed method is verified through exhaustive testing  on 234 document images.<\/p>\n<\/p><\/div>\n<\/td>\n<\/tr>\n<tr class=\"talk\">\n<td class=\"time\">11:45<\/td>\n<td> <a name=\"talk:30521\"\/> <\/p>\n<div class=\"authors\"> <a name=\"talk:30521\"\/> <a href=\"http:\/\/easychair.org\/smart-program\/DocEng2016\/person83.html\" class=\"person\">Emilio Granell<\/a> and <a class=\"person\" href=\"http:\/\/easychair.org\/smart-program\/DocEng2016\/person84.html\">Carlos David Martinez Hinarejos<\/a> <\/div>\n<div class=\"title\">A Multimodal Crowdsourcing Framework for Transcribing Historical Handwritten Documents<\/div>\n<div class=\"speaker\"\/>\n<div class=\"abstract\">\n<p>ABSTRACT.  Transcription of handwritten historical documents is one of the main topics in document analysis systems, due to humanistic and cultural reasons. State-of-the-art handwritten text recognition systems allow to speed up the transcription task. Currently, this automatic transcription  is far from being perfect, and human expert revision is required in order to obtain the actual transcription. In this context, crowdsourcing  emerged as a powerful tool for massive transcription at a relatively low cost, since the supervision effort of professional transcribers may be dramatically reduced. However, current transcription crowdsourcing platforms are mainly limited to the use of non-mobile devices, since the  use of keyboards in mobile devices is not friendly enough for most users. This work presents the alternative of using speech dictation of handwritten text lines as transcription source in a crowdsourcing platform. The experiments explore how an initial handwritten text recognition hypothesis can be improved by using the contribution of speech recognition from several speakers, providing as a final result a better hypothesis to be amended by a professional transcriber with less effort.<\/p>\n<\/p><\/div>\n<\/td>\n<\/tr>\n<tr class=\"talk\">\n<td class=\"time\">12:15<\/td>\n<td> <a name=\"talk:30522\"\/> <\/p>\n<div class=\"authors\"> <a name=\"talk:30522\"\/> <a class=\"person\" href=\"http:\/\/easychair.org\/smart-program\/DocEng2016\/person79.html\">Lucia Noce<\/a>, <a class=\"person\" href=\"http:\/\/easychair.org\/smart-program\/DocEng2016\/person80.html\">Ignazio Gallo<\/a>, <a href=\"http:\/\/easychair.org\/smart-program\/DocEng2016\/person81.html\" class=\"person\">Alessandro Zamberletti<\/a> and <a class=\"person\" href=\"http:\/\/easychair.org\/smart-program\/DocEng2016\/person82.html\">Alessandro Calefati<\/a> <\/div>\n<div class=\"title\">Embedded Textual Content for Document Image Classification with Convolutional Neural Networks<\/div>\n<div class=\"speaker\"\/>\n<div class=\"abstract\">\n<p>ABSTRACT.  In this paper we introduce a novel document image classification method  based on combined visual and textual information. The proposed algorithm&#8217;s pipeline is inspired to the ones of other recent state-of-the-art methods which perform document image classification using Convolutional Neural Networks. The main addition of our work is the introduction of a preprocessing step embedding additional textual information into the processed document images. To do so we combine Object Character Recognition and Natural Language Processing algorithms to extract and manipulate relevant text concepts from document images. Such textual information is then visually embedded within each document image to improve the classification results of a Convolutional Neural Network. Our experiments prove that the overall document classification accuracy of a Convolutional Neural Network trained using these text-augmented document images is considerably higher than the one achieved by a similar model trained solely on classic document images, especially when  different classes of documents share similar visual characteristics.<\/p>\n<\/p><\/div>\n<\/td>\n<\/tr>\n<\/tbody>\n<\/table><\/div>\n<div class=\"session notalk\"> <a name=\"session:10878\"> <\/p>\n<div class=\"heading\"> <span class=\"interval\">12:45-13:00<\/span> <span class=\"title\">Session 17: Closing Notes<\/span> <\/div>\n<p> <\/a> <\/div>\n<div class=\"session\">\n<div class=\"lunchbreak\"> <span class=\"interval\">13:00-14:30<\/span> <span class=\"title\">Lunch Break<\/span> <\/div>\n<\/p><\/div>\n<\/p><\/div>\n<\/td>\n<\/tr>\n<\/tbody>\n<\/table><\/div>\n","protected":false},"excerpt":{"rendered":"<p>&nbsp; &nbsp; Overview| Wednesday| Thursday| Friday 09:30-10:45 Session 15: Applications &amp; Security Chair: David F. Brailsford 09:30 Rodrigo Laiola Guimaraes,<\/p>\n","protected":false},"author":6,"featured_media":0,"parent":0,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"","meta":[],"_links":{"self":[{"href":"https:\/\/doceng2016.cvl.tuwien.ac.at\/index.php?rest_route=\/wp\/v2\/pages\/599"}],"collection":[{"href":"https:\/\/doceng2016.cvl.tuwien.ac.at\/index.php?rest_route=\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/doceng2016.cvl.tuwien.ac.at\/index.php?rest_route=\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/doceng2016.cvl.tuwien.ac.at\/index.php?rest_route=\/wp\/v2\/users\/6"}],"replies":[{"embeddable":true,"href":"https:\/\/doceng2016.cvl.tuwien.ac.at\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=599"}],"version-history":[{"count":3,"href":"https:\/\/doceng2016.cvl.tuwien.ac.at\/index.php?rest_route=\/wp\/v2\/pages\/599\/revisions"}],"predecessor-version":[{"id":637,"href":"https:\/\/doceng2016.cvl.tuwien.ac.at\/index.php?rest_route=\/wp\/v2\/pages\/599\/revisions\/637"}],"wp:attachment":[{"href":"https:\/\/doceng2016.cvl.tuwien.ac.at\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=599"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}