D
Dejan Pecevski
Researcher at Graz University of Technology
Publications - 14
Citations - 2370
Dejan Pecevski is an academic researcher from Graz University of Technology. The author has contributed to research in topics: Artificial neural network & Python (programming language). The author has an hindex of 12, co-authored 14 publications receiving 2139 citations. Previous affiliations of Dejan Pecevski include University of Graz.
Papers
More filters
Journal ArticleDOI
Simulation of networks of spiking neurons: A review of tools and strategies
Romain Brette,Michelle Rudolph,Ted Carnevale,Michael L. Hines,David Beeman,James M. Bower,Markus Diesmann,Markus Diesmann,Abigail Morrison,Philip H. Goodman,Frederick C. Harris,Milind Zirpe,Thomas Natschläger,Dejan Pecevski,G. Bard Ermentrout,Mikael Djurfeldt,Anders Lansner,Olivier Rochel,Thierry Viéville,Eilif Muller,Andrew P. Davison,Sami El Boustani,Alain Destexhe +22 more
TL;DR: In this paper, a review of different aspects of the simulation of spiking neural networks is presented, with the aim of identifying the appropriate integration strategy and simulation tool to use for a given modeling problem related to spiking networks.
Journal ArticleDOI
PyNN: A Common Interface for Neuronal Network Simulators.
Andrew P. Davison,Daniel Brüderle,Jochen Martin Eppler,Jens Kremkow,Eilif Muller,Dejan Pecevski,Laurent Perrinet,Pierre Yger +7 more
TL;DR: PyNN increases the productivity of neuronal network modelling by providing high-level abstraction, by promoting code sharing and reuse, and by providing a foundation for simulator-agnostic analysis, visualization and data-management tools.
Journal ArticleDOI
A Learning Theory for Reward-Modulated Spike-Timing-Dependent Plasticity with Application to Biofeedback
TL;DR: The resulting learning theory predicts that even difficult credit-assignment problems can be solved in a self-organizing manner through reward-modulated STDP, and provides a possible functional explanation for trial-to-trial variability, which is characteristic for cortical networks of neurons but has no analogue in currently existing artificial computing systems.
Journal ArticleDOI
Probabilistic Inference in General Graphical Models through Sampling in Stochastic Networks of Spiking Neurons
TL;DR: Ubiquitous stochastic features of networks of spiking neurons, such as trial-to-trial variability and spontaneous activity, are necessary ingredients of the underlying computational organization and can be scaled up to neural emulations of probabilistic inference in fairly large graphical models.
Journal ArticleDOI
PCSIM: A Parallel Simulation Environment for Neural Circuits Fully Integrated with Python
TL;DR: This paper investigates how the automatically generated bidirectional interface and PCSIM's object-oriented modular framework enable the user to adopt a hybrid modeling approach: using and extending PC SIM's functionality either employing pure Python or C++ and thus combining the advantages of both worlds.