1 code implementation • 30 May 2021 • Aristeidis Panos, Ioannis Kosmidis, Petros Dellaportas
We adopt the interpretability offered by a parametric, Hawkes-process-inspired conditional probability mass function for the marks and apply variational inference techniques to derive a general and scalable inferential framework for marked point processes.
1 code implementation • 11 Jan 2020 • Ioannis Kosmidis, Nicola Lunardon
That penalized objective relates closely to information criteria for model selection, and can be further enhanced with plug-in penalties to deliver reduced-bias $M$-estimates with extra properties, like finiteness in models for categorical data.
Methodology Statistics Theory Statistics Theory 62F10, 62F12, 62J12
1 code implementation • 5 Dec 2018 • Ioannis Kosmidis, David Firth
This paper studies the finiteness properties of the reduced-bias estimator for logistic regression that results from penalization of the likelihood by Jeffreys' invariant prior; and it provides geometric insights on the shrinkage towards equiprobability that the penalty induces.
Statistics Theory Methodology Statistics Theory 62J12, 62F10, 62F12, 62F03
1 code implementation • 29 Oct 2018 • Heather L. Turner, Jacob van Etten, David Firth, Ioannis Kosmidis
This means that the worth of each item can always be estimated by maximum likelihood, with finite standard error.
Computation
no code implementations • 4 Jul 2018 • Alkeos Tsokos, Santhosh Narayanan, Ioannis Kosmidis, Gianluca Baio, Mihai Cucuringu, Gavin Whitaker, Franz J. Király
The parameters of the Bradley-Terry extensions are estimated by maximizing the log-likelihood, or an appropriately penalized version of it, while the posterior densities of the parameters of the hierarchical Poisson log-linear model are approximated using integrated nested Laplace approximations.
no code implementations • 11 Apr 2018 • Ioannis Kosmidis, Euloge Clovis Kenne Pagui, Nicola Sartori
This paper presents an integrated framework for estimation and inference from generalized linear models using adjusted score equations that result in mean and median bias reduction.
Methodology 62J12, 62F03, 62F12
1 code implementation • 22 Feb 2018 • Thomas E. Bartlett, Ioannis Kosmidis, Ricardo Silva
Separation of these two types of sparsity is achieved with the introduction of a novel prior structure, which draws on ideas from the Bayesian lasso and from copula modelling.
Methodology