Department of Computer Science and Engineering, Jahangirnagar University, Savar, Dhaka.
Department of Data Science and Engineering, Bangabandhu Sheikh Mujibur Rahman Digital University, Gazipur, Bangladesh.
PLoS One. 2024 Oct 10;19(10):e0311161. doi: 10.1371/journal.pone.0311161. eCollection 2024.
People use search engines for various topics and items, from daily essentials to more aspirational and specialized objects. Therefore, search engines have taken over as people's preferred resource. The "How To" prefix has become familiar and widely used in various search styles to find solutions to particular problems. This search allows people to find sequential instructions by providing detailed guidelines to accomplish specific tasks. Categorizing instructional text is also essential for task-oriented learning and creating knowledge bases. This study uses the "How To" articles to determine the multi-label instruction category. We have brought this work with a dataset comprising 11,121 observations from wikiHow, where each record has multiple categories. To find out the multi-label category meticulously, we employ some transformer-based deep neural architectures, such as Generalized Autoregressive Pretraining for Language Understanding (XLNet), Bidirectional Encoder Representation from Transformers (BERT), etc. In our multi-label instruction classification process, we have reckoned our proposed architectures using accuracy and macro f1-score as the performance metrics. This thorough evaluation showed us much about our strategy's strengths and drawbacks. Specifically, our implementation of the XLNet architecture has demonstrated unprecedented performance, achieving an accuracy of 97.30% and micro and macro average scores of 89.02% and 93%, a noteworthy accomplishment in multi-label classification. This high level of accuracy and macro average score is a testament to the effectiveness of the XLNet architecture in our proposed 'InstructNet' approach. By employing a multi-level strategy in our evaluation process, we have gained a more comprehensive knowledge of the effectiveness of our proposed architectures and identified areas for forthcoming improvement and refinement.
人们使用搜索引擎来查询各种主题和项目,从日常用品到更有抱负和专业的物品。因此,搜索引擎已经取代了人们作为首选的资源。“如何”前缀已经在各种搜索方式中变得熟悉和广泛使用,以找到解决特定问题的方法。这种搜索允许人们通过提供完成特定任务的详细指导来找到顺序指令。对教学文本进行分类对于面向任务的学习和创建知识库也是至关重要的。本研究使用“如何”文章来确定多标签指令类别。我们带来了这项工作,其中包含了来自 wikiHow 的 11121 个观察结果的数据集,其中每个记录都有多个类别。为了细致地找出多标签类别,我们采用了一些基于转换器的深度神经网络架构,如 XLNet(Generalized Autoregressive Pretraining for Language Understanding)、BERT(Bidirectional Encoder Representation from Transformers)等。在我们的多标签指令分类过程中,我们使用准确性和宏 F1 分数作为性能指标来评估我们提出的架构。这项全面的评估让我们对我们策略的优缺点有了更深入的了解。具体来说,我们对 XLNet 架构的实现展示了前所未有的性能,达到了 97.30%的准确性,以及微和宏平均分数 89.02%和 93%,这在多标签分类中是一个值得注意的成就。这种高精度和宏平均分数证明了 XLNet 架构在我们提出的“InstructNet”方法中的有效性。通过在评估过程中采用多层次策略,我们对我们提出的架构的有效性有了更全面的了解,并确定了即将改进和完善的领域。