Showing 1 - 20 results of 23 for search 'neural explicit coding', query time: 0.09s Refine Results
  1. 1

    Feature dependence graph based source code loophole detection method by Hongyu YANG, Haiyun YANG, Liang ZHANG, Xiang CHENG

    Published 2023-01-01
    “…Given the problem that the existing source code loophole detection methods did not explicitly maintain the semantic information related to the loophole in the source code, which led to the difficulty of feature extraction of loo-phole statements and the high false positive rate of loophole detection, a source code loophole detection method based on feature dependency graph was proposed.First, extracted the candidate loophole statements in the function slice, and gen-erated the feature dependency graph by analyzing the control dependency chain and data dependency chain of the candi-date loophole statements.Secondly, the word vector model was used to generate the initial node representation vector of the feature dependency graph.Finally, a loophole detection neural network oriented to feature dependence graph was constructed, in which the graph learning network learned the heterogeneous neighbor node information of the feature de-pendency graph and the detection network extracted global features and performed loophole detection.The experimental results show that the recall rate and F1 score of the proposed method are improved by 1.50%~22.32% and 1.86%~16.69% respectively, which is superior to the existing method.…”
    Get full text
    Article
  2. 2

    Feature dependence graph based source code loophole detection method by Hongyu YANG, Haiyun YANG, Liang ZHANG, Xiang CHENG

    Published 2023-01-01
    “…Given the problem that the existing source code loophole detection methods did not explicitly maintain the semantic information related to the loophole in the source code, which led to the difficulty of feature extraction of loo-phole statements and the high false positive rate of loophole detection, a source code loophole detection method based on feature dependency graph was proposed.First, extracted the candidate loophole statements in the function slice, and gen-erated the feature dependency graph by analyzing the control dependency chain and data dependency chain of the candi-date loophole statements.Secondly, the word vector model was used to generate the initial node representation vector of the feature dependency graph.Finally, a loophole detection neural network oriented to feature dependence graph was constructed, in which the graph learning network learned the heterogeneous neighbor node information of the feature de-pendency graph and the detection network extracted global features and performed loophole detection.The experimental results show that the recall rate and F1 score of the proposed method are improved by 1.50%~22.32% and 1.86%~16.69% respectively, which is superior to the existing method.…”
    Get full text
    Article
  3. 3
  4. 4

    Relating sparse and predictive coding to divisive normalization. by Yanbo Lian, Anthony N Burkitt

    Published 2025-05-01
    “…In this paper, we show how sparse coding, predictive coding, and divisive normalization can be described within a unified framework, and illustrate this explicitly within the context of a two-layer neural learning model of sparse/predictive coding. …”
    Get full text
    Article
  5. 5

    Scale Is Not All You Need: Revisiting the Biomimetic Roots of Deep Learning to Overcome Fundamental Limitations by Malek Wahidi, Anthony Rizk, Rodrigue Imad

    Published 2025-01-01
    “…Specifically, we explore the emerging topics of Liquid Neural Networks, recurrent vision architectures, and Predictive Coding as representative examples of such solutions. …”
    Get full text
    Article
  6. 6

    Dance as mindful movement: a perspective from motor learning and predictive coding by W. Tecumseh Fitch, Rebecca Barnstaple

    Published 2024-11-01
    “…However, to make this conception explicit and testable, we need an empirically verifiable characterization of “mindful movement.” …”
    Get full text
    Article
  7. 7
  8. 8

    Disentangling signal and noise in neural responses through generative modeling. by Kendrick Kay, Jacob S Prince, Thomas Gebhart, Greta Tuckute, Jingyang Zhou, Thomas Naselaris, Heiko H Schütt

    Published 2025-07-01
    “…We end by discussing other situations that may benefit from GSN's characterization of signal and noise, such as estimation of noise ceilings for computational models of neural activity. A code toolbox for GSN is provided with both MATLAB and Python implementations.…”
    Get full text
    Article
  9. 9
  10. 10

    Regression analysis and artificial neural networks for predicting pine species volume in community forests by Wenceslao Santiago-García

    Published 2025-11-01
    “…These variables were explicitly incorporated into the ANN structure, enabling the model to learn complex, non-linear interactions. …”
    Get full text
    Article
  11. 11

    Architecture-aware minimization (A2M): how to find flat minima in neural architecture search by Matteo Gambella, Fabrizio Pittorino, Manuel Roveri

    Published 2025-01-01
    “…Neural architecture search (NAS) has become an essential tool for designing effective and efficient neural networks. …”
    Get full text
    Article
  12. 12

    A neural network-based method for input parameter optimization of edge transport modeling utilizing experimental diagnostics by Y. Luo, S. Xu, Y. Liang, E. Wang, J. Cai, Y. Feng, D. Reiter, A. Knieps, S. Brezinsek, D. Harting, M. Krychowiak, D. Gradic, P. Ren, D. Zhang, Y. Gao, G. Fuchert, A. Pandey, M. Jakubowski, the W7-X Team

    Published 2025-01-01
    “…EMC3-EIRENE simulations using the maximum a posteriori estimates derived from these posterior distributions reproduce experimental measurements with satisfactory accuracy. This neural network-based method significantly reduces computational costs and the need for manual parameter tuning, and it can be generalized to other similar modeling codes.…”
    Get full text
    Article
  13. 13

    Modelling the future of cleaner energy: Explainable artificial intelligence model for green hydrogen production rate estimation by Okorie Ekwe Agwu, Saad Alatefi, Ahmad Alkouh

    Published 2025-07-01
    “…Overall, the model can be used in two modes: online, by integrating it into software programs, and offline, by simply entering parameter values into the explicit model without having to run long lines of code.…”
    Get full text
    Article
  14. 14

    A Semantically Enhanced Label Prediction Method for Imbalanced POI Data Category Distribution by Hongwei Zhang, Qingyun Du, Shuai Zhang, Renfei Yang

    Published 2024-10-01
    “…Furthermore, most neural network classification models implicitly learn the semantic knowledge of different categories from training datasets, neglecting the explicit semantic information offered by natural language labels. …”
    Get full text
    Article
  15. 15

    Suboptimal but intact integration of Bayesian components during perceptual decision-making in autism by Laurina Fazioli, Bat-Sheva Hadad, Rachel N. Denison, Amit Yashar

    Published 2025-01-01
    “…This hypothesis has gained momentum in recent years, partly because it can be implemented both at the computational level, as in Bayesian perception, and at the level of canonical neural microcircuitry, as in predictive coding. However, empirical investigations have yielded conflicting results with evidence remaining limited. …”
    Get full text
    Article
  16. 16

    Visuospatial sequence learning without seeing. by Clive R Rosenthal, Christopher Kennard, David Soto

    Published 2010-07-01
    “…Learning sequential associations between non-adjacent visual stimuli (higher-order visuospatial dependencies) can occur either with or without awareness (explicit vs. implicit learning) of the products of learning. …”
    Get full text
    Article
  17. 17

    FMRI reveals a dissociation between grasping and perceiving the size of real 3D objects. by Cristiana Cavina-Pratesi, Melvyn A Goodale, Jody C Culham

    Published 2007-05-01
    “…The demonstration of dual coding of an object for the purposes of action on the one hand and perception on the other in the same healthy brains offers a substantial contribution to the current debate about the nature of the neural coding that takes place in the dorsal and ventral streams.…”
    Get full text
    Article
  18. 18

    Generating diversity and securing completeness in algorithmic retrosynthesis by Florian Mrugalla, Christopher Franz, Yannic Alber, Georg Mogk, Martín Villalba, Thomas Mrziglod, Kevin Schewior

    Published 2025-05-01
    “…Scientific Contribution: We adapt Depth-First Proof-Number Search (DFPN) (Please refer to https://github.com/Bayer-Group/bayer-retrosynthesis-search for the accompanying source code.) and its variants, which have been applied to retrosynthesis before, to produce a set of solutions, with an explicit focus on diversity. …”
    Get full text
    Article
  19. 19

    Large language models for depression recognition in spoken language integrating psychological knowledge by Yupei Li, Shuaijie Shao, Manuel Milling, Manuel Milling, Björn W. Schuller, Björn W. Schuller, Björn W. Schuller, Björn W. Schuller

    Published 2025-08-01
    “…Depression is a growing concern gaining attention in both public discourse and AI research. While deep neural networks (DNNs) have been used for its recognition, they still lack real-world effectiveness. …”
    Get full text
    Article
  20. 20

    RTide: Automating the Tidal Response Method by Thomas Monahan, Tianning Tang, Stephen Roberts, Thomas A. A. Adcock

    Published 2025-06-01
    “…Our approach embeds a class of neural networks capable of representing any arbitrary Volterra series—the mathematical basis of the response method—within the classic framework. …”
    Get full text
    Article