Skip to content

Commit 4f6e9bb

Browse files
committed
Update publications
1 parent 0ca8057 commit 4f6e9bb

24 files changed

Lines changed: 500 additions & 3 deletions

File tree

_people/samuel-bell.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -10,6 +10,7 @@ scholar: yfgSAi8AAAAJ
1010
github: samueljamesbell
1111
linkedin: samueljamesbell
1212
start: 2018-10-01
13+
end: 2023-04-19
1314
crsid: sjb326
1415
supervisor: ndl21
1516
position: PhD Student

_people/vidhi-lalchand.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -5,6 +5,7 @@ family: Lalchand
55
crsid: vr308
66
student: true
77
start: 2022-02-07
8+
end: 2023-10-01
89
website: https://www.vidhilalchand.co.uk
910
orcid: null
1011
position: Research Associate

_projects/autoai.md

Lines changed: 6 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -9,13 +9,16 @@ featured_image: alina-grubnyak-ziqkhi7417a-unsplash.jpg
99
people:
1010
- andrei-paleyes
1111
- christian-cabrera
12-
- eric-meissner
1312
- jessica-montgomery
14-
- mala-virdee
15-
- markus-kaiser
1613
- morine-amutorine
1714
- neil-d-lawrence
15+
- diana-robinson
16+
alumni:
17+
- eric-meissner
18+
- mala-virdee
19+
- markus-kaiser
1820
- pierre-thodoroff
21+
1922
publications:
2023
- an-empirical-evaluation-of-flow-based-programming-in-the-machine-learning-deployment-context
2124
- benchmarking-real-time-reinforcement-learning
Lines changed: 26 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,26 @@
1+
---
2+
layout: techreport
3+
title: "Inconsistency in Conference Peer Review: Revisiting the 2014 NeurIPS Experiment"
4+
date: 2021-09-20
5+
published: 2021-09-20
6+
abstract: |
7+
In this paper we revisit the 2014 NeurIPS experiment that examined inconsistency in conference peer review. We determine that 50% of the variation in reviewer quality scores was subjective in origin. Further, with seven years passing since the experiment we find that for *accepted* papers, there is no correlation between quality scores and impact of the paper as measured as a function of citation count. We trace the fate of rejected papers, recovering where these papers were eventually published. For these papers we find a correlation between quality scores and impact. We conclude that the reviewing process for the 2014 conference was good for identifying poor papers, but poor for identifying good papers. We give some suggestions for improving the reviewing process but also warn against removing the subjective element. Finally, we suggest that the real conclusion of the experiment is that the community should place less onus on the notion of 'top-tier conference publications' when assessing the quality of individual researchers. For NeurIPS 2021, the PCs are repeating the experiment, as well as conducting new ones.
8+
author:
9+
- given: Corinna
10+
family: Cortes
11+
- given: Neil D.
12+
family: Lawrence
13+
arxiv: 2109.09774
14+
website: https://arxiv.org/abs/2109.09774
15+
doi: 10.48550/arXiv.2109.09774
16+
subjects:
17+
- Digital Libraries (cs.DL)
18+
- Machine Learning (cs.LG)
19+
submission_history:
20+
- version: v1
21+
date: 2021-09-20
22+
size: 2,012 KB
23+
submitter: Neil Lawrence
24+
email: view email
25+
software: https://github.com/lawrennd/neurips2014/
26+
---

_publications/2021-10-08-differentially-private-regression-and-classificaiton-with-sparse-gaussian-processes.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -18,6 +18,7 @@ firstpage: 1
1818
lastpage: 41
1919
key: Smith-differentially21
2020
doi:
21+
sscholar: 202712625
2122
html: https://jmlr.org/papers/v22/19-017.html
2223
pdf: https://jmlr.org/papers/volume22/19-017/19-017.pdf
2324
software: https://github.com/lionfish0/dp4gp
Lines changed: 31 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,31 @@
1+
---
2+
layout: article
3+
title: "Challenges in Machine Learning Deployment: A Survey of Case Studies"
4+
abstract: In recent years, machine learning has transitioned from a field of
5+
academic research interest to a field capable of solving real-world business
6+
problems. However, the deployment of machine learning models in production
7+
systems can present a number of issues and concerns. This survey reviews
8+
published reports of deploying machine learning solutions in a variety of use
9+
cases, industries and applications and extracts practical considerations
10+
corresponding to stages of the machine learning deployment workflow. By
11+
mapping found challenges to the steps of the machine learning deployment
12+
workflow we show that practitioners face issues at each stage of the
13+
deployment process. The goal of this paper is to lay out a research agenda to
14+
explore approaches addressing these challenges.
15+
published: 2022-04-30
16+
author:
17+
- given: Andrei
18+
family: Paleyes
19+
mlatcl_title: andrei-paleyes
20+
person_page: andrei-paleyes
21+
- given: Raoul-Gabriel
22+
family: Urma
23+
- given: Neil D.
24+
family: Lawrence
25+
person_page: neil-d-lawrence
26+
journal: ACM Comput. Surv.
27+
publisher: Association for Computing Machinery
28+
address: New York, NY, USA
29+
website: https://doi.org/10.1145/3533378
30+
arxiv: 2011.09926
31+
---
Lines changed: 45 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,45 @@
1+
---
2+
layout: techreport
3+
title: "Modelling Technical and Biological Effects in scRNA-seq Data with Scalable GPLVMs"
4+
date: 2022-09-14
5+
published: 2022-11-06
6+
abstract: |
7+
Single-cell RNA-seq datasets are growing in size and complexity, enabling the study of cellular composition changes in various biological/clinical contexts. Scalable dimensionality reduction techniques are in need to disentangle biological variation in them, while accounting for technical and biological confounders. In this work, we extend a popular approach for probabilistic non-linear dimensionality reduction, the Gaussian process latent variable model, to scale to massive single-cell datasets while explicitly accounting for technical and biological confounders. The key idea is to use an augmented kernel which preserves the factorisability of the lower bound allowing for fast stochastic variational inference. We demonstrate its ability to reconstruct latent signatures of innate immunity recovered in Kumasaka et al. (2021) with 9x lower training time. We further analyze a COVID dataset and demonstrate across a cohort of 130 individuals, that this framework enables data integration while capturing interpretable signatures of infection. Specifically, we explore COVID severity as a latent dimension to refine patient stratification and capture disease-specific gene expression.
8+
author:
9+
- given: Vidhi
10+
family: Lalchand
11+
- given: Aditya
12+
family: Ravuri
13+
- given: Emma
14+
family: Dann
15+
- given: Natsuhiko
16+
family: Kumasaka
17+
- given: Dinithi
18+
family: Sumanaweera
19+
- given: Rik G.H.
20+
family: Lindeboom
21+
- given: Shaista
22+
family: Madad
23+
- given: Sarah A.
24+
family: Teichmann
25+
- given: Neil D.
26+
family: Lawrence
27+
arxiv: 2209.06716
28+
website: https://arxiv.org/abs/2209.06716
29+
doi: 10.48550/arXiv.2209.06716
30+
subjects:
31+
- Machine Learning (cs.LG)
32+
- Genomics (q-bio.GN)
33+
- Applications (stat.AP)
34+
- Machine Learning (stat.ML)
35+
submission_history:
36+
- version: v1
37+
date: 2022-09-14
38+
size: 12,682 KB
39+
submitter: Vidhi Lalchand Miss
40+
email: view email
41+
- version: v2
42+
date: 2022-11-06
43+
size: 12,682 KB
44+
comments: Machine Learning and Computational Biology Symposium (Oral), 2022
45+
---
Lines changed: 22 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,22 @@
1+
---
2+
layout: techreport
3+
title: "AI for Science: an emerging agenda"
4+
date: 2023-03-08
5+
published: 2023-03-08
6+
abstract: |
7+
This report documents the programme and the outcomes of Dagstuhl Seminar 22382 "Machine Learning for Science: Bridging Data-Driven and Mechanistic Modelling". Today's scientific challenges are characterised by complexity. Interconnected natural, technological, and human systems are influenced by forces acting across time- and spatial-scales, resulting in complex interactions and emergent behaviours. Understanding these phenomena -- and leveraging scientific advances to deliver innovative solutions to improve society's health, wealth, and well-being -- requires new ways of analysing complex systems. The transformative potential of AI stems from its widespread applicability across disciplines, and will only be achieved through integration across research domains. AI for science is a rendezvous point. It brings together expertise from AI and application domains; combines modelling knowledge with engineering know-how; and relies on collaboration across disciplines and between humans and machines. Alongside technical advances, the next wave of progress in the field will come from building a community of machine learning researchers, domain experts, citizen scientists, and engineers working together to design and deploy effective AI tools. This report summarises the discussions from the seminar and provides a roadmap to suggest how different communities can collaborate to deliver a new wave of progress in AI and its application for scientific discovery.
8+
author:
9+
- given: Philipp
10+
family: Berens
11+
- given: Kyle
12+
family: Cranmer
13+
- given: Neil D.
14+
family: Lawrence
15+
- given: Ulrike
16+
family: von Luxburg
17+
- given: Jessica
18+
family: Montgomery
19+
arxiv: 2303.04217
20+
website: https://arxiv.org/abs/2303.04217
21+
doi: 10.48550/arXiv.2303.04217
22+
---
Lines changed: 19 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,19 @@
1+
---
2+
layout: inproceedings
3+
title: "Dataflow graphs as complete causal graphs"
4+
published: 2023
5+
author:
6+
- given: Andrei
7+
family: Paleyes
8+
- given: Siyuan
9+
family: Guo
10+
- given: Bernhard
11+
family: Schölkopf
12+
- given: Neil D.
13+
family: Lawrence
14+
abstract: |
15+
Component-based development is one of the core principles behind modern software engineering practices. Understanding of causal relationships between components of a software system can yield significant benefits to developers. Yet modern software design approaches make it difficult to track and discover such relationships at system scale, which leads to growing intellectual debt. In this paper we consider an alternative approach to software design, flow-based programming (FBP), and draw the attention of the community to the connection between dataflow graphs produced by FBP and structural causal models. With expository examples we show how this connection can be leveraged to improve day-to-day tasks in software projects, including fault localisation, business analysis and experimentation.
16+
booktitle: 2023 IEEE/ACM 2nd International Conference on AI Engineering–Software Engineering Approaches
17+
website: https://arxiv.org/abs/2303.09552
18+
arxiv: 2303.09552
19+
---
Lines changed: 18 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,18 @@
1+
---
2+
layout: techreport
3+
title: "Dimensionality Reduction as Probabilistic Inference"
4+
published: 2023-04-15
5+
abstract: |
6+
Dimensionality reduction (DR) algorithms compress high-dimensional data into a lower dimensional representation while preserving important features of the data. DR is a critical step in many analysis pipelines as it enables visualisation, noise reduction and efficient downstream processing of the data. In this work, we introduce the ProbDR variational framework, which interprets a wide range of classical DR algorithms as probabilistic inference algorithms in this framework. ProbDR encompasses PCA, CMDS, LLE, LE, MVU, diffusion maps, kPCA, Isomap, (t-)SNE, and UMAP. In our framework, a low-dimensional latent variable is used to construct a covariance, precision, or a graph Laplacian matrix, which can be used as part of a generative model for the data. Inference is done by optimizing an evidence lower bound. We demonstrate the internal consistency of our framework and show that it enables the use of probabilistic programming languages (PPLs) for DR. Additionally, we illustrate that the framework facilitates reasoning about unseen data and argue that our generative models approximate Gaussian processes (GPs) on manifolds. By providing a unified view of DR, our framework facilitates communication, reasoning about uncertainties, model composition, and extensions, particularly when domain knowledge is present.
7+
author:
8+
- given: Aditya
9+
family: Ravuri
10+
- given: Francisco
11+
family: Vargas
12+
- given: Vidhi
13+
family: Lalchand
14+
- given: Neil D.
15+
family: Lawrence
16+
website: https://arxiv.org/abs/2304.07658
17+
arxiv: 2304.07658
18+
---

0 commit comments

Comments
 (0)