共找到 20 条结果
Understanding historical datasets, such as the England and Wales infant mortality data, for local government districts can provide valuable insights into our changing society. Such analyses can prove challenging in practice, due to frequent changes in the boundaries of local government districts for which records are collected. One solution adopted in the literature to overcome such practical challenges is to pre-process data using areal interpolation to render the units consistent over the time period of focus. However, such methods are prone to errors. In this paper we introduce a novel changepoint method to detect instances where interpolation performs poorly. We demonstrate the utility of our method on original data, and also demonstrate how correcting interpolation errors can affect the clustering of the infant mortality curves.
We propose a new method of determining regional and city boundaries based on the Valeriepieris circle, the smallest circle containing a given fraction of the data. By varying the fraction in the circle we can map complex spatial data to a simple model of concentric rings which we then fit to determine natural density cutoffs. We apply this method to population, occupation, economic and transport data from England and Wales, finding that the regions determined by this method affirm well known social facts such as the disproportionate wealth of London or the relative isolation of the North East and South West of England. We then show how different data sets give us different views of the same cities, providing insight into their development and dynamics.
We introduce Oxford Day-and-Night, a large-scale, egocentric dataset for novel view synthesis (NVS) and visual relocalisation under challenging lighting conditions. Existing datasets often lack crucial combinations of features such as ground-truth 3D geometry, wide-ranging lighting variation, and full 6DoF motion. Oxford Day-and-Night addresses these gaps by leveraging Meta ARIA glasses to capture egocentric video and applying multi-session SLAM to estimate camera poses, reconstruct 3D point clouds, and align sequences captured under varying lighting conditions, including both day and night. The dataset spans over 30 $\mathrm{km}$ of recorded trajectories and covers an area of 40,000 $\mathrm{m}^2$, offering a rich foundation for egocentric 3D vision research. It supports two core benchmarks, NVS and relocalisation, providing a unique platform for evaluating models in realistic and diverse environments.
Human Immunodeficiency Virus (HIV) has posed a major global health challenge for decades, and forecasting HIV diagnoses continues to be a critical area of research. However, capturing the complex spatial and temporal dependencies of HIV transmission remains challenging. Conventional Message Passing Neural Network (MPNN) models rely on a fixed binary adjacency matrix that only encodes geographic adjacency, which is unable to represent interactions between non-contiguous counties. Our study proposes a deep learning architecture Mobility-Aware Transformer-Message Passing Neural Network (MAT-MPNN) framework to predict county-level HIV diagnosis rates across California, Florida, and the New England region. The model combines temporal features extracted by a Transformer encoder with spatial relationships captured through a Mobility Graph Generator (MGG). The MGG improves conventional adjacency matrices by combining geographic and demographic information. Compared with the best-performing hybrid baseline, the Transformer MPNN model, MAT-MPNN reduced the Mean Squared Prediction Error (MSPE) by 27.9% in Florida, 39.1% in California, and 12.5% in New England, and improved the Predictive Mode
This chapter examines the challenges of the revised opt out system and the secondary use of health data in England. The analysis of this data could be very valuable for science and medical treatment as well as for the discovery of new drugs. For this reason, the UK government established the care.data program in 2013. The aim of the project was to build a central nationwide database for research and policy planning. However, the processing of personal data was planned without proper public engagement. Research has suggested that IT companies, such as in the Google DeepMind deal case, had access to other kinds of sensitive data and failed to comply with data protection law. Since May 2018, the government has launched the national data opt out system with the hope of regaining public trust. Nevertheless, there are no evidence of significant changes in the ND opt out, compared to the previous opt out system. Neither in the use of secondary data, nor in the choices that patients can make. The only notorious difference seems to be in the way that these options are communicated and framed to the patients. Most importantly, according to the new ND opt out, the type 1 opt out option, which
We propose a new approach to identifying geographical clustering and hotspots of inequality from decadal census data. We use diffusion mapping to study the 181,408 Output Areas in England and Wales, which allows us to decompose the feature structures of countries in the census data space. Additionally, we develop a new localization metric inspired by statistical physics to uncover the importance of minority groups in London. The results of our study can be applied to other census-like data constructions that include spatial localization and differentiation from low degrees of freedom. This new approach can help us better understand the patterns of social deprivation and segregation across the country and aid in the development of policies to address these issues.
This paper introduces a large-scale multi-modal dataset captured in and around well-known landmarks in Oxford using a custom-built multi-sensor perception unit as well as a millimetre-accurate map from a Terrestrial LiDAR Scanner (TLS). The perception unit includes three synchronised global shutter colour cameras, an automotive 3D LiDAR scanner, and an inertial sensor - all precisely calibrated. We also establish benchmarks for tasks involving localisation, reconstruction, and novel-view synthesis, which enable the evaluation of Simultaneous Localisation and Mapping (SLAM) methods, Structure-from-Motion (SfM) and Multi-view Stereo (MVS) methods as well as radiance field methods such as Neural Radiance Fields (NeRF) and 3D Gaussian Splatting. To evaluate 3D reconstruction the TLS 3D models are used as ground truth. Localisation ground truth is computed by registering the mobile LiDAR scans to the TLS 3D models. Radiance field methods are evaluated not only with poses sampled from the input trajectory, but also from viewpoints that are from trajectories which are distant from the training poses. Our evaluation demonstrates a key limitation of state-of-the-art radiance field methods:
There is a growing academic interest as well as commercial exploitation of millimetre-wave scanning radar for autonomous vehicle localisation and scene understanding. Although several datasets to support this research area have been released, they are primarily focused on urban or semi-urban environments. Nevertheless, rugged offroad deployments are important application areas which also present unique challenges and opportunities for this sensor technology. Therefore, the Oxford Offroad Radar Dataset (OORD) presents data collected in the rugged Scottish highlands in extreme weather. The radar data we offer to the community are accompanied by GPS/INS reference - to further stimulate research in radar place recognition. In total we release over 90GiB of radar scans as well as GPS and IMU readings by driving a diverse set of four routes over 11 forays, totalling approximately 154km of rugged driving. This is an area increasingly explored in literature, and we therefore present and release examples of recent open-sourced radar place recognition systems and their performance on our dataset. This includes a learned neural network, the weights of which we also release. The data and tools
In England, it is anecdotally remarked that the number of Greggs bakeries to be found in a town is a reliable measure of the area's 'Northern-ness'. Conversely, a commercial competitor to Greggs in the baked goods and sandwiches market, Pret-a-Manger, is reputed to be popular in more 'southern' areas of England. Using a Support Vector Machine and an Artificial Neural Network (ANN) Regression Model, the relative geographical distributions of Greggs and Pret have been utilised for the first time to quantify the North-South divide in England. The calculated dividing lines were each compared to another line, based on Gross Domestic Household Income (GDHI). The lines match remarkably well, and we conclude that this is likely because much of England's wealth is concentrated in London, as are most of England's Pret-a-Manger shops. Further studies were conducted based on the relative geographical distributions of popular supermarkets Morrisons and Waitrose, which are also considered to have a North-South association. This analysis yields different results. For all metrics, the North-South dividing line passes close to the M1 Watford Gap services. As a common British idiom, this location is
The prevalence of dementia is set to explode throughout the 21st century. This trend has already started in developed countries and will continue to place heavy pressures on both public health and social care services across the world. No cure for dementia is likely within the foreseeable future, however, medical research highlights the potential to diminish the risk of dementia onset. Over one-third of dementia cases may be preventable if certain risk factors are addressed at the individual, clinical, and population level. This research further explores these modifiable risk factors and quantifies areal risk through the use of a composite index. The index operates at National Health Service Clinical Commission Group level to assess spatial differences across England. Clear spatial patterns are observed between the north and south of the country, and between London and the remainder of the country. The framework adopted in this research provides a firm foundation upon which similar indices could be produced, potentially at finer spatial resolutions, incorporating more informed local knowledge and data on relevant dementia risk factors.
The COVID-19 pandemic has had high mortality rates in the elderly and frail worldwide, particularly in care homes. This is driven by the difficulty of isolating care homes from the wider community, the large population sizes within care facilities (relative to typical households), and the age/frailty of the residents. To quantify the mortality risk posed by disease, the case fatality risk (CFR) is an important tool. This quantifies the proportion of cases that result in death. Throughout the pandemic, CFR amongst care home residents in England has been monitored closely. To estimate CFR, we apply both novel and existing methods to data on deaths in care homes, collected by Public Health England and the Care Quality Commission. We compare these different methods, evaluating their relative strengths and weaknesses. Using these methods, we estimate temporal trends in the instantaneous CFR (at both daily and weekly resolutions) and the overall CFR across the whole of England, and dis-aggregated at regional level. We also investigate how the CFR varies based on age and on the type of care required, dis-aggregating by whether care homes include nursing staff and by age of residents. This
This extended abstract was written to accompany an invited talk at the 2021 SC-Square Workshop, where the author was asked to give an overview of SC-Square progress to date. The author first reminds the reader of the definition of SC-Square, then briefly outlines some of the history, before picking out some (personal) scientific highlights.
Oxford-style debating is a well-known tool in social sciences. Such formal discussions on particular topics are widely used by historians and sociologists. However, when we try to go beyond standard thinking, it turns out that Oxford-style debating can be a great educational tool in telecommunication and computer science. This article presents this unusual method of education at technical universities and in the IT industry, and describes its features and challenges. Best practices and examples of debating are provided, taking into account emerging topics in telecommunications and computer science, such as cybersecurity. The article also contains feedback from IT engineers who participated in Oxford-style debates. All this aims to encourage this form of education in telecommunication and computer science.
This paper builds and extends on the authors' previous work related to the algorithmic tool, Cylindrical Algebraic Decomposition (CAD), and one of its core applications, Real Quantifier Elimination (QE). These topics are at the heart of symbolic computation and were first implemented in computer algebra systems decades ago, but have recently received renewed interest as part of the ongoing development of SMT solvers for non-linear real arithmetic. First, we consider the use of iterated univariate resultants in traditional CAD, and how this leads to inefficiencies, especially in the case of an input with multiple equational constraints. We reproduce the workshop paper [Davenport and England, 2023], adding important clarifications to our suggestions first made there to make use of multivariate resultants in the projection phase of CAD. We then consider an alternative approach to this problem first documented in [McCallum and Brown, 2009] which redefines the actual object under construction, albeit only in the case of two equational constraints. We correct an unhelpful typo and provide a proof missing from that paper. We finish by revising the topic of how to deal with SMT or Real QE
There has been an increasing number of applications of machine learning to the field of Computer Algebra in recent years, including to the prominent sub-field of Symbolic Integration. However, machine learning models require an abundance of data for them to be successful and there exist few benchmarks on the scale required. While methods to generate new data already exist, they are flawed in several ways which may lead to bias in machine learning models trained upon them. In this paper, we describe how to use the Risch Algorithm for symbolic integration to create a dataset of elementary integrable expressions. Further, we show that data generated this way alleviates some of the flaws found in earlier methods.
String Field Theory is a formulation of String Theory as a Quantum Field Theory in target space. It allows to tame the infrared divergences of String Theory and to approach its non-perturbative structure and background independence. This article gives a concise overview on the subject and of some of the main recent progress. Note: Review article for Oxford Research Encyclopedia of Physics
The increased integration of variable renewable generation into the power systems, along with the phase-out of fossil-based power stations, necessitate procuring more flexibility from the demand sectors. The electrification of the residential heat sector is an option to decarbonise the heat sector in the United Kingdom. The inherent flexibility that is available in the residential heat sector, in the form of the thermal inertia of buildings, is expected to play an important role in supporting the critical task of short-term balancing of electricity supply and demand. This paper proposes a method for characterising the locally aggregated flexibility envelope from the electrified residential heat sector, considering the most influential factors including outdoor and indoor temperature, thermal mass and heat loss of dwellings. Applying the method to England and Wales as a case study, demonstrated a significant potential for a temporary reduction of electricity demand for heating even during cold weather. Total electricity demand reductions of approximately 25 GW to 85 GW were shown to be achievable for the outdoor temperature of 10 degreeC and -5 degreeC, respectively. Improving the e
In May 2022, a cluster of mpox cases were detected in the UK that could not be traced to recent travel history from an endemic region. Over the coming months, the outbreak grew, with over 3000 total cases reported in the UK, and similar outbreaks occurring worldwide. These outbreaks appeared linked to sexual contact networks between gay, bisexual and other men who have sex with men. Following the COVID-19 pandemic, local health systems were strained, and therefore effective surveillance for mpox was essential for managing public health policy. However, the mpox outbreak in the UK was characterised by substantial delays in the reporting of the symptom onset date and specimen collection date for confirmed positive cases. These delays led to substantial backfilling in the epidemic curve, making it challenging to interpret the epidemic trajectory in real-time. Many nowcasting models exist to tackle this challenge in epidemiological data, but these lacked sufficient flexibility. We have developed a novel nowcasting model using generalised additive models to correct the mpox epidemic curve in England, and provide real-time characteristics of the state of the epidemic, including the real-
Cylindrical Algebraic Decomposition (CAD) by projection and lifting requires many iterated univariate resultants. It has been observed that these often factor, but to date this has not been used to optimise implementations of CAD. We continue the investigation into such factorisations, writing in the specific context of SC-Square.
Einstein's blackboard is a well-known exhibit at the History of Science Museum at Oxford University. However, it is much less well known that the writing on the board provides a neat summary of a work of historic importance, Einstein's 1931 model of the expanding universe. As a visual representation of one of the earliest models of the universe to be proposed in the wake of Hubble's observations of the nebulae, the blackboard provides an intriguing snapshot of a key moment in modern astronomy and cosmology. In addition, one line on the blackboard that is not in Einstein's 1931 paper casts useful light on some anomalies in the calculations of that paper.