https://ph01.tci-thaijo.org/index.php/ecticit/issue/feed ECTI Transactions on Computer and Information Technology (ECTI-CIT) 2026-04-23T09:41:58+07:00 Prof.Dr.Prabhas Chongstitvattana and Prof.Dr.Chidchanok Lursinsap chief.editor.cit@gmail.com Open Journal Systems <p style="text-align: justify;">ECTI Transactions on Computer and Information Technology (ECTI-CIT) is published by the Electrical Engineering/Electronics, Computer, Telecommunications and Information Technology (ECTI) Association which is a professional society that aims to promote the communication between electrical engineers, computer scientists, and IT professionals. Contributed papers must be original that advance the state-of-the-art applications of Computer and Information Technology. Both theoretical contributions (including new techniques, concepts, and analyses) and practical contributions (including system experiments and prototypes, and new applications) are encouraged. The submitted manuscript must have not been copyrighted, published, submitted, or accepted for publication elsewhere. This journal employs <em><strong>a double-blind review</strong></em>, which means that throughout the review process, the identities of both the reviewer and the author are concealed from each other. The manuscript text should not contain any commercial references, such as<span class="L57vkdwH4 ZIjt03VBzHWC"> company names</span>, university names, trademarks, commercial acronyms, or part numbers. The manuscript length must be at least 8 pages and no longer than 10 pages with two (2) columns.</p> <p style="text-align: justify;"><strong>Journal Abbreviation</strong>: ECTI-CIT</p> <p style="text-align: justify;"><strong>Since</strong>: 2005</p> <p style="text-align: justify;"><strong>ISSN</strong>: 2286-9131 (Online)</p> <p style="text-align: justify;"><strong>Language</strong>: English</p> <p style="text-align: justify;"><strong>Review Method</strong>: Double Blind</p> <p style="text-align: justify;"><strong>Issues Per Year</strong>: 2 Issues (from 2005-2020), 3 Issues (in 2021), and 4 Issues (from 2022).</p> <p style="text-align: justify;"><strong>Publication Fee</strong>: Free of charge.</p> <p style="text-align: justify;"><strong>Published Articles</strong>: Review Article / Research Article / Invited Article (only for an invitation provided by editors)</p> <p style="text-align: justify;"><strong>Scopus preview:</strong> https://www.scopus.com/sourceid/21100899864</p> <p style="text-align: justify;"><strong>DOI prefix for the ECTI Transactions</strong> is: 10.37936/ (https://doi.org/)</p> https://ph01.tci-thaijo.org/index.php/ecticit/article/view/262099 Intelligent Honeypot for Web Applications: Leveraging Seq2Seq and Reinforcement Learning 2026-02-05T15:45:34+07:00 Ananya Varadarajan pes2202100749@pesu.pes.edu Ashwin Chandrasekaran pes2202100832@pesu.pes.edu Rachana Binumohan pes2202100742@pesu.pes.edu Rahul Huliyar Ravishankar pes2202100378@pesu.pes.edu Gokul Kannan Sadasivam gokul@pes.edu <p><span style="font-weight: 400;">An intelligent honeypot system designed to mimic legitimate websites using Sequence-to-Sequence (Seq2Seq) learning and Deep Q-Learning. The system generates realistic, contextually appropriate responses to attacker queries, prolonging interactions and providing insights into malicious behaviors while safeguarding actual systems. The Seq2Seq model, trained on HTTP request-response pairs, enables the honeypot to produce responses that closely resemble those of real servers, enhancing its ability to deceive attackers. Deep Q-Learning optimizes engagement by selecting the most effective responses through a custom reward function, balancing realism and interactivity to maximize session length. Performance was evaluated using metrics such as Response Realism Rate (RRR), Semantic Consistency Accuracy (SCA), and Average Session Length (ASL). The honeypot achieved an RRR of 92.3%, an SCA of 89.7%, and a 94.5% Optimal Response Selection Rate (ORSR). These advancements increased ASL by 143.5%, from 3.2 to 7.8 exchanges, reflecting prolonged attacker engagement. By integrating Seq2Seq and Deep Q-Learning, this honeypot demonstrates significant improvements in generating convincing responses and sustaining interactions. These results contribute to modern cybersecurity by providing a practical and theoretical framework for developing next-generation honeypots capable of deceiving attackers and gathering actionable intelligence.</span></p> 2026-02-28T00:00:00+07:00 Copyright (c) 2026 ECTI Transactions on Computer and Information Technology (ECTI-CIT) https://ph01.tci-thaijo.org/index.php/ecticit/article/view/263972 YUV-based Deep Learning Super-Resolution for Bitrate Reduction and ROI Preservation in Modern Video Codecs 2026-02-12T13:34:51+07:00 Lertluck Leela-amornsin l.lertluck@gmail.com Nuttapon Vanakittistien nuttapon.vana@gmail.com Nattee Niparnan nattee@gmail.com Pitchaya Sitthi-amorn pitchaya@cp.eng.chula.ac.th Attawith Sudsang attawith@cp.eng.chula.ac.th <p><span style="font-weight: 400;">High Efficiency Video Coding (HEVC) and its successors, such as Versatile Video Coding (VVC), offer substantial bitrate reductions, yet challenges remain in preserving visual fidelity under bandwidth and computational constraints. This paper proposes a deep learning-based super-resolution (SR) framework that operates natively in the YUV color space, eliminating costly RGB-YUV conversions and integrating seamlessly with modern video compression pipelines. We develop two convolutional network architectures trained on YUV-formatted video data: a full 3-channel model and a lightweight two-stream variant that separately processes luminance (Y) and chrominance (UV) channels using compact subnetworks. The proposed method enhances both full-frame and region-of-interest (ROI) quality, outperforming conventional HEVC baselines in terms of rate-distortion efficiency. Evaluations on diverse video sequences demonstrate significant bitrate savings and effective ROI preservation, with the lightweight model offering a practical solution for AI-driven applications in resource-constrained environments.</span></p> 2026-03-07T00:00:00+07:00 Copyright (c) 2026 ECTI Transactions on Computer and Information Technology (ECTI-CIT) https://ph01.tci-thaijo.org/index.php/ecticit/article/view/264201 Hybrid WangchanBERTa Architectures for Multi-Class Thai Sentiment Analysis 2026-03-12T10:15:14+07:00 Panida Songram panida.s@msu.ac.th Suchart Khummanee suchart.k@msu.ac.th Khanabhorn Kawattikul khanabhorn_ka@rmutto.ac.th Nittaya Muangnak nittaya.mu@ku.th <p class="Bodytext"><span style="font-weight: 400;">The rapid growth of the restaurant industry in Thailand has intensified the importance of online reviews, which significantly shape customer perceptions and influence business performance. Sentiment analysis has emerged as an effective computational approach for extracting customer opinions from such reviews; however, multi-class sentiment classification in Thai remains challenging due to the language's non-segmented structure and the issue of class imbalance. This study investigates three hybrid deep learning modelsWangchanBERTa-MLP, WangchanBERTa- CNN, and WangchanBERTa-BiLSTMby integrating WangchanBERTa, a Thai-specific pre-trained language model, with different neural architectures. Using a balanced dataset of restaurant reviews obtained through SMOTE, the models were evaluated based on accuracy, precision, recall, and F1-score. The experimental results show that WangchanBERTa- BiLSTM performed the best overall, achieving an accuracy of 85.22% and significantly improving the classification of neutral and positive sentiments compared to the BERT-based models and other hybrid methods.</span></p> 2026-03-21T00:00:00+07:00 Copyright (c) 2026 ECTI Transactions on Computer and Information Technology (ECTI-CIT) https://ph01.tci-thaijo.org/index.php/ecticit/article/view/264238 Comparison of CNN Architectures for Thai Medicinal Plant Classification 2026-03-12T11:07:51+07:00 Sompong Valuvanathorn sompong.v@ubu.ac.th Chanchai Supaartagorn chanchai.s@ubu.ac.th <p><span style="font-weight: 400;">Thai medicinal plants are essential to traditional healthcare and local livelihoods. However, many Thai medicinal plants have similar morphological characteristics such as shape, colour, and texture. This problem leads to misidentification and misclassification. Image classifiers utilizing convolutional neural networks (CNNs), which are a class of deep learning models, provide a scalable substitute for manual classification. This study aims to evaluate and compare the performance of three CNN architectures (DenseNet-121, EfficientNet-B3, and MobileNetV2) for classifying 10 species of Thai medicinal plants. The dataset comprises 5,000 leaf images representing 10 species (500 images per species). This study partitioned the dataset into 80% training set and a 20% test set. To enhance model generalization, we applied data augmentation techniques-specifically rotation, flipping, and colour manipulation. Furthermore, we utilized TensorFlow and Keras on Google Colab with GPU acceleration to train the models. Evaluation metrics include accuracy, precision, recall, F1 score, model size, inference time, and CPU utilization. The results highlight a trade-off between accuracy and efficiency: DenseNet-121 achieved the highest accuracy at 96.0% and a Matthews Correlation Coefficient (MCC) of 0.9558. Statistical analysis confirmed that DenseNet-121 significantly outperformed the other architectures (p &lt; 0.05), albeit with a higher inference time (579.22 s). Notably, EfficientNet-B3 and MobileNetV2 both achieved an accuracy of 93.4%, with MobileNetV2 performing the best in terms of model size (11.07 MB) and inference time (3.86 s). In conclusion, DenseNet-121 is the most accurate model, while MobileNetV2 is best suited for real-time applications due to its lightweight and rapid inference time. EfficientNet-B3 offers an optimal balance between accuracy and computational efficiency.</span></p> 2026-03-28T00:00:00+07:00 Copyright (c) 2026 ECTI Transactions on Computer and Information Technology (ECTI-CIT) https://ph01.tci-thaijo.org/index.php/ecticit/article/view/265498 Enhancing LTE Handover Decision using Optimised Extreme Gradient Boosting and Rule-Based Decision-Support 2026-02-12T13:42:03+07:00 Noormadinah Allias noormadinah@yahoo.com Megat Norulazmi Megat Mohamed Noor megatnorulazmi@unikl.edu.my Mohd Nazri Ismail m.nazri@upnm.edu.my Mohd Taha Ismail mtaha@unikl.edu.my <p>Long-Term Evolution (LTE) provides low-latency, high-data-rate services, which are essential for delay-sensitive applications such as video streaming and online gaming. Despite this, user mobility among cells can degrade network performance, so efficient handover management is crucial to maintain Quality of Service (QoS). Traditional handover mechanisms use static control parameters, such as hysteresis margin and time-to-trigger, that are not flexible for working with users' dynamic mobility or a range of user trajectories. In this paper, we present a learning-based optimised data-driven approach for LTE handover decision support. An XGBoost model trained with Hyperopt to learn the relationship between user movement angle and handover performance parameters. Interpretable if-then rules are developed to modify the handover control parameters adaptively. Experimental results further show that the performance of the fixed-parameter solutions depends on the maximum handover delay and the mean time to handover, including the minimum handover rate, indicating that a single configuration is unlikely to provide the best performance across all mobility scenarios. The solution offers an efficient, scalable, and interpretable decision-support system to improve LTE handover efficiency in dynamic wireless networks.</p> 2026-03-28T00:00:00+07:00 Copyright (c) 2026 ECTI Transactions on Computer and Information Technology (ECTI-CIT) https://ph01.tci-thaijo.org/index.php/ecticit/article/view/264970 VhAR-Net: A Cross-Modal Representation Learning Framework for Text-Based Vehicle Retrieval 2026-03-05T05:58:44+07:00 Lin Qian 6772100069@stu.pim.ac.th JIAN QU jianqu@pim.ac.th <p><span style="font-weight: 400;">With the advancement of intelligent transportation systems and large-scale urban video surveillance technologies, vehicle image retrieval based on textual descriptions has become increasingly important. Although MCANet exhibits effectiveness in multi-scale feature alignment, significant limitations persist in accurate fine-grained semantic matching. To address this, we present VhAR-Neta modular cross-modal retrieval architecture that enables independent design and flexible combination of feature interaction mechanisms, enhancement strategies, and supervisory signals. This framework employs ResNet-50 as the visual representation extractor and BERT as the textual semantic encoder, and innovatively introduces a cross-modal attention unit that establishes explicit associations between image representations and linguistic descriptions. Concurrently, the system establishes small-object aware auxiliary supervision through binary classification tasks targeting discriminative fine-grained vocabulary, thereby directing the network toward distinctive microscopic semantic units. Empirical evaluations on the public benchmark dataset T2I-VeRi indicate that the optimal configuration achieves 75% Top-1 retrieval accuracy, with Top-10 recall covering 85% of relevant samples. After introducing the feature enhancement module and small-object supervision mechanism, the cumulative matching rate for Top-5 and Top-10 both reached 85%, demonstrating improved retrieval robustness.</span></p> 2026-04-04T00:00:00+07:00 Copyright (c) 2026 ECTI Transactions on Computer and Information Technology (ECTI-CIT) https://ph01.tci-thaijo.org/index.php/ecticit/article/view/265040 A Dynamic Memory Routing Framework for Multimodal Conversational Question Answering in Large Language Models 2026-03-26T13:33:38+07:00 Walisa Romsaiyud walisar@gmail.com <p><span style="font-weight: 400;">Multimodal Question Answering (MQA) in large language models (LLMs) requires adaptive modeling of modality relevance across conversational turns. However, existing approaches rely on static fusion strategies that treat modalities uniformly and fail to capture dynamic modality importance. To address this limitation, we propose DynaRoute, a dynamic memory routing framework for LLM-based MQA. DynaRoute integrates a Bi-LSTM-based conversational memory to model evolving dialogue context and a query-conditioned routing mechanism that dynamically assigns modality relevance at each interaction step. The resulting representations are processed by an LLM-based decoder to generate context-aware responses. Experiments on four benchmarks-VQA-v2, GQA, VisDial, and A-OKVQA-demonstrate consistent improvements over unimodal, static fusion, and mixture-of-experts baselines. DynaRoute achieves an improvement of up to 8.7% under noisy conditions and 6.5% under clean settings, while also obtaining the highest multi-turn consistency (68.9) and robustness (81.3) scores. These results highlight the effectiveness of memory-aware dynamic routing and establish DynaRoute as a principled framework for conversational multimodal question answering.</span></p> 2026-04-25T00:00:00+07:00 Copyright (c) 2026 ECTI Transactions on Computer and Information Technology (ECTI-CIT) https://ph01.tci-thaijo.org/index.php/ecticit/article/view/264823 Flip-Robust Neural Image Assessment (FR-NIMA) for Spatially Consistent IQA 2026-03-12T21:28:25+07:00 Rawesak Tanawongsuwan rawesak.tan@mahidol.ac.th Sukanya Phongsuphap sukanya.pho@mahidol.ac.th <p><span style="font-weight: 400;">Neural Image Assessment (NIMA) has become a widely adopted approach for blind image quality assessment (BIQA), yet it remains sensitive to simple spatial transformations such as horizontal ips. Such variation can lead to inconsistent predictions, even when the perceived visual content remains largely unchanged. To address this issue, we introduce Flip-Robust Neural Image Assessment (FR-NIMA), an enhanced training strategy that enhances the spatial robustness of BIQA models. Instead of modifying network architectures, FR-NIMA incorporates a flip-consistency regularization term that penalizes discrepancies between the predicted quality distributions of an image and its horizontally flipped counterpart. Two variants-one-branch and two-branch formulations-are explored, both introducing no additional model parameters. FR-NIMA is evaluated across four CNN backbones (MobileNetV2, VGG19, Xception, InceptionV3) and one Vision Transformer (ViT-Small) using the LIVE dataset and two additional test sets representing distinct scene types. Performance is assessed using complementary metrics, including the Test Loss (EMD2), Absolute Flip Gap (|FlipGap|), Flip-Consistency Win Rate (FCWR), Average Flip- Gap Delta (AFGD), and Average Flip-Gap Ratio (AFGR). Experimental results demonstrate that FR-NIMA effectively reduces ip-gap magnitude and variability while maintaining comparable test accuracy across all back- bones. These findings establish FR-NIMA as a simple yet effective framework for enhancing the stability, spatial consistency, and trustworthiness of deep IQA models.</span></p> 2026-04-25T00:00:00+07:00 Copyright (c) 2026 ECTI Transactions on Computer and Information Technology (ECTI-CIT) https://ph01.tci-thaijo.org/index.php/ecticit/article/view/263898 A Kernel-Aware Framework for Energy-Efficient FPGA Edge Detection Using XNOR-popcount and Selective Approximate Arithmetic 2026-04-23T09:21:42+07:00 Van-Khoa Pham khoapv@hcmute.edu.vn <p><span style="font-weight: 400;">Edge-vision systems on resource-constrained platforms require low-latency and energy-efficient front-end processing. Conventional gradient-based edge operators continue to rely on multiply-accumulate (MAC) operations, which can increase logic utilization, switching activity, and power consumption in eld-programmable gate array (FPGA) implementations. This study introduces a kernel-aware arithmetic-selection framework, supported by synthesized evidence, for energy-efficient FPGA-based edge detection. The framework combines XNORpopcount-based arithmetic mapping with selective approximate computation. Its central design principle is to replace MAC operations with exclusive-NOR-population count (XNORpopcount) when the kernel structure is suitable for binary-friendly arithmetic, and to apply approximation only to the remaining adder-dominated stages when a complete XNOR-popcount mapping is not practical. Under this rule, Prewitt-like operators are mapped to an XNORpopcount datapath over their active nonzero taps. In contrast, Sobel-like operators are realized using a hybrid datapath that combines binary matching, shift-add processing, and approximate accumulation. The resulting framework shows that multiplier removal is the dominant source of hardware savings, while approximate arithmetic provides a controlled secondary optimization. Overall, the proposed approach establishes a structured design methodology for low-power FPGA edge-detection architectures on embedded platforms.</span></p> 2026-04-30T00:00:00+07:00 Copyright (c) 2026 ECTI Transactions on Computer and Information Technology (ECTI-CIT) https://ph01.tci-thaijo.org/index.php/ecticit/article/view/265497 Machine Learning-based Estimation of Foot Parameters for Custom Insole Production: A Comprehensive Analysis of Structural Measurements 2026-03-12T10:08:28+07:00 Wudhichart Sawangphol wudhichart.saw@mahidol.ac.th Jidapa Kraisangka jidapa.kra@mahidol.ac.th Pisit Praiwattana pisit.pra@mahidol.ac.th Pilailuck Panphattarasap pilailuck.pan@mahidol.ac.th <p><span style="font-weight: 400;">Precise quantification of foot morphology is critical for clinical diagnostics, biomechanics, and personalized orthotic design. Conventional anthropometric methods often remain labor-intensive or dependent on specialized hardware, necessitating more efficient predictive frameworks. This study develops and validates a series of machine learning (ML) models to predict nine essential foot anthropometric parameters. Leveraging a dataset of 544 independent foot samples from 272 participants, encompassing high, normal, and flat arch types. The models were evaluated using a 10-fold cross-validation strategy to ensure robust generalizability. Our pipeline integrates correlation-based feature selection with hyperparameter-optimized regression algorithms, including XGBoost, Random Forest, Support Vector Regressors, Neural Networks, and Linear Regression. The results demonstrate high predictive fidelity, with Mean Absolute Errors (MAE) consistently remaining below 0.5 cm. This level of precision meets the 0.5 cm clinical tolerance threshold established through expert consultation for in-sole production, while simultaneously aligning with international footwear sizing increments, thereby confirming the framework's practical utility in real-world manufacturing. Although parameters such as the length from heel to midfoot area and the length from heel to distal metatarsal head achieved exceptional precision (MAE of 0.012 cm and 0.026 cm, respectively), predicting arch height remains a notable challenge. This research underscores the necessity of optimal feature engineering and algorithm selection in automating foot morphometric assessment.</span></p> 2026-04-30T00:00:00+07:00 Copyright (c) 2026 ECTI Transactions on Computer and Information Technology (ECTI-CIT) https://ph01.tci-thaijo.org/index.php/ecticit/article/view/264177 FEDIS: Graph-Based Social Media Evidence Collection with Correlation-Aware Forensic Analysis 2026-04-23T09:41:58+07:00 Somchart Fugkeaw somchart@siit.tu.ac.th Pratuck Khubpatiwitthayakul pratuck.khub@gmail.com Pongsapon Ateetanan pongsapon.ateeta@gmail.com Diyorbek Bakhtiyor Ugli Okyulov diyorbekokyulov@gmail.com <p><span style="font-weight: 400;">Online Social Networks (OSNs) have emerged as a major source of digital evidence in cybercrime investigations, abuse detection, and online incident analysis. Publicly available data, such as posts, comments, reactions, and user interactions, provide critical insights into suspicious activities and behavioral patterns. However, extracting and analyzing such data in a forensic context remains challenging. Existing social media data acquisition approaches primarily rely on web scraping or API-based techniques that produce unstructured outputs (e.g., raw text and images) without preserving the relationships among entities such as users, posts, and interactions. This results in the loss of contextual information essential for forensic analysis. Furthermore, data collection and analysis are often performed separately, resulting in delayed investigations and limited real-time insight. To address these limitations, this paper focuses on two key forensic requirements: (i) efficient and structured acquisition of social media evidence and (ii) correlation-aware analysis of interactions. We propose FEDIS, a unified forensic data acquisition and analysis system that integrates a hybrid DOM-TAO data model, graph-based representation, and parallel keyword-based search. Experimental results demonstrate that FEDIS achieves complete relationship preservation, improves data collection completeness, and significantly reduces search latency compared to traditional approaches, making it practical for real-world social media forensic investigations.</span></p> 2026-04-30T00:00:00+07:00 Copyright (c) 2026 ECTI Transactions on Computer and Information Technology (ECTI-CIT) https://ph01.tci-thaijo.org/index.php/ecticit/article/view/265851 Multi-Speaker Thai Speech Synthesis Using Transfer Learning 2026-04-23T09:38:24+07:00 Kittikan Charoenrattana ckittikan@kkumail.com Pusadee Seresangtakul pusadee@kku.ac.th Pongsathon Janyoi pongsathon@kku.ac.th <p>This paper investigates transfer learning for developing a multi-speaker Thai text-to-speech (TTS) system under low-resource conditions, with a focus on cross-lingual knowledge transfer from source languages with different phonological and prosodic characteristics. The model is pre-trained on speech data from Thai, Mandarin Chinese, and English, and subsequently fine-tuned using Thai speech from three target speakers, each with only 1530 minutes of data, covering both male and female speakers. Both objective and subjective evaluations consistently demonstrate that Thai pre-training achieves the best overall performance. Among the cross-lingual models, transfer learning from Mandarin Chinese outperforms transfer from English, yielding lower F0 RMSE (20.79 vs. 21.65) and higher MOS scores (3.55 vs. 3.44). In addition, speaker-dependent analysis indicates that speaker gender has a noticeable influence on synthesis quality under limited data conditions, suggesting that acoustic similarity between pre-training data and target speakers can affect the effectiveness of knowledge transfer. However, this factor is secondary to linguistic and prosodic similarity. Natural speech achieves a MOS of 4.93, while the Thai, Mandarin Chinese, and English-pre-trained models obtain MOS scores of 3.65, 3.55, and 3.44, respectively. These results highlight the importance of linguistic proximity in cross-lingual multi-speaker TTS, particularly tonal and prosodic similarity between source and target languages. Overall, the study confirms that transfer learning is an effective approach for low-resource Thai TTS and that tonal source languages provide more beneficial knowledge transfer than non-tonal, accent-based languages.</p> 2026-04-30T00:00:00+07:00 Copyright (c) 2026 ECTI Transactions on Computer and Information Technology (ECTI-CIT) https://ph01.tci-thaijo.org/index.php/ecticit/article/view/264386 LLM-Driven Annotation: A Scalable Framework for Automated Multi-Label Coffee Flavor Classification 2026-03-19T08:11:40+07:00 Prasara Jakkaew prasara.jak@mfu.ac.th Nacha Chondamrongkul nacha.cho@mfu.ac.th Paweena Suebsombut paweena.sue@mfu.ac.th <p><span style="font-weight: 400;">The subjective and unstructured nature of coffee tasting notes creates a significant data annotation bottleneck, limiting the application of computational methods in sensory science. This study introduces a novel end-to-end framework for automating multi-label coffee flavor classification, integrating LLM-driven annotation with transformer-based classification and local interpretability analysis. Google's Gemini-1.5-Flash was employed to perform evidence-based annotation on 8,327 expert evaluations authored by certified Q Graders, generating a high-quality training dataset across 17 flavor categories. Critically, the Q Grader provenance of the source texts enables a reverse-mapping validation framework: statistically significant and directionally coherent correlations between LLM-derived labels and expert quantitative scores (Floral r = 0.32, Roasted r = −0.25, all p &lt;0.001) provide implicit ground truth without requiring separate annotation effort. Reliability analysis revealed substantial consistency for concrete physical descriptors (mean κ = 0.68) but notably lower agreement for abstract sensory concepts such as Mouthfeel (κ = 0.10), identifying two distinct reliability regimes that define the framework's operational boundaries. A ne-tuned BERT model trained on these annotations outperformed a TF- IDF baseline, achieving a Micro F1-score of 0.9164 versus 0.8763 and a Hamming Loss of 0.0640 versus 0.0969. To address severe class imbalance (up to 125×), Focal Loss (γ = 2.0) combined with per-class threshold optimization successfully recovered detection of rare defect categories, improving Macro-F1 from 0.656 to 0.719. Local interpretability analysis via LIME further confirmed that model predictions align with domain-expert sensory reasoning. These results demonstrate that an LLM-driven annotation pipeline offers a scalable, transparent, and effective solution to the data bottleneck in sensory science, establishing a robust methodological foundation for interpretable classification across other sensory-driven domains.</span></p> 2026-04-30T00:00:00+07:00 Copyright (c) 2026 ECTI Transactions on Computer and Information Technology (ECTI-CIT)