We use cookies to distinguish you from other users and to provide you with a better experience on our websites. Close this message to accept cookies or find out how to manage your cookie settings.
To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
The data volumes generated by theWALLABY atomic Hydrogen (HI) survey using the Australian Square Kilometre Array Pathfinder (ASKAP) necessitate greater automation and reliable automation in the task of source-finding and cataloguing. To this end, we introduce and explore a novel deep learning framework for detecting low Signal-to-Noise Ratio (SNR) HI sources in an automated fashion. Specifically, our proposed method provides an automated process for separating true HI detections from false positives when used in combination with the Source Finding Application (SoFiA) output candidate catalogues. Leveraging the spatial and depth capabilities of 3D ConvolutionalNeuralNetworks (CNNs), our method is specifically designed to recognize patterns and features in three-dimensional space, making it uniquely suited for rejecting false positive sources in low SNR scenarios generated by conventional linear methods. As a result, our approach is significantly more accurate in source detection and results in considerably fewer false detections compared to previous linear statistics-based source finding algorithms. Performance tests using mock galaxies injected into real ASKAP data cubes reveal our method’s capability to achieve near-100% completeness and reliability at a relatively low integrated SNR ∼ 3 – 5. An at-scale version of this tool will greatly maximise the science output from the upcoming widefield HI surveys.
We have conducted a widefield, wideband, snapshot survey using the Australian SKA Pathfinder (ASKAP) referred to as the Rapid ASKAP Continuum Survey (RACS). RACS covers ≈ 90% of the sky, with multiple observing epochs in three frequency bands sampling the ASKAP frequency range of 700–1 800 MHz. This paper describes the third major epoch at 1 655.5 MHz, RACS-high, and the subsequent imaging and catalogue data release. The RACS-high observations at 1 655.5MHz are otherwise similar to the previously released RACS-mid (at 1367.5 MHz), and were calibrated and imaged with minimal changes. From the 1 493 images covering the sky up to declination ≈ +48°, we present a catalogue of 2 677 509 radio sources. The catalogue is constructed from images with a median root-mean-square noise of ≈ 195 μJy PSF−1 (point-spread function) and a median angular resolution of 11″. 8 × 8″. 1. The overall reliability of the catalogue is estimated to be 99.18 %, and we find a decrease in reliability as angular resolution improves. We estimate the brightness scale to be accurate to 10 %, and the astrometric accuracy to be within ≈ 0″. 6 in right ascension and ≈ 0″. 7 in declination after correction of a systematic declination-dependent offset. All data products from RACS-high, including calibrated visibility datasets, images from individual observations, full-sensitivity mosaics, and the all-sky catalogue are available at the CSIRO ASKAP Science Data Archive.
Researchers regularly use large survey studies to examine public political opinion. Surveys running over days and months will necessarily incorporate religious occasions that can introduce variation in public opinion. Using recent survey data from Israel, this study demonstrates that giving surveys on religious occasions (e.g., the Sabbath, Hannukah, Sukkot) can elicit different opinion responses. These effects are found among both religious and non-religious respondents. While incorporating these fluctuations is realistic in longer-term surveys, surveys fielded in a short window inadvertently drawing heavily on a holiday or holy day sample may bias their findings. This study thus urges researchers to be cognizant of ambient religious context when conducting survey studies.
We examine the energy distribution of the fast radio burst (FRB) population using a well-defined sample of 63 FRBs from the Australian Square Kilometre Array Pathfinder (ASKAP) radio telescope, 28 of which are localised to a host galaxy. We apply the luminosity-volume ($V/V_{\mathrm{max}}$) test to examine the distribution of these transient sources, accounting for cosmological and instrumental effects, and determine the energy distribution for the sampled population over the redshift range $0.01 \lesssim z \lesssim 1.02$. We find the distribution between $10^{23}$ and $10^{26}$ J Hz$^{-1}$ to be consistent with both a pure power-law with differential slope $\gamma=-1.96 \pm 0.15$, and a Schechter function with $\gamma = -1.82 \pm 0.12$ and downturn energy $E_\mathrm{max} \sim 6.3 \, \times 10^{25}$ J Hz$^{-1}$. We identify systematic effects which currently limit our ability to probe the luminosity function outside this range and give a prescription for their treatment. Finally, we find that with the current dataset, we are unable to distinguish between the evolutionary and spectral models considered in this work.
The explosion of attention to measuring and understanding implicit bias has been influential inside and outside the academy. The purpose of this chapter is to balance the conversation about how to unpack and understand implicit bias, with an exploration of what we know about Whites’ explicit bias, and how surveys and other data can be used to measure it. This chapter begins with a review of survey-based data on White racial attitudes that reveal complex trends and patterns, with some topics showing changes for the better, but others showing persistent negative or stagnant trends. Drawing on examples using a variety of methodological tools, including (1) traditional survey questions; (2) survey-based mode/question wording experiments; (3) open-ended questions embedded in surveys; and (4) in-depth interviews, I illustrate what explicit racial biases can look like, and how they might be consequential. I argue that a full understanding of intergroup relations requires sophisticated methods and theories surrounding both explicit and implicit biases, how they function separately and in combination, and their causes and consequences.
When surveys contain direct questions about sensitive topics, participants may not provide their true answers. Indirect question techniques incentivize truthful answers by concealing participants’ responses in various ways. The Crosswise Model aims to do this by pairing a sensitive target item with a non-sensitive baseline item, and only asking participants to indicate whether their responses to the two items are the same or different. Selection of the baseline item is crucial to guarantee participants’ perceived and actual privacy and to enable reliable estimates of the sensitive trait. This research makes the following contributions. First, it describes an integrated methodology to select the baseline item, based on conceptual and statistical considerations. The resulting methodology distinguishes four statistical models. Second, it proposes novel Bayesian estimation methods to implement these models. Third, it shows that the new models introduced here improve efficiency over common applications of the Crosswise Model and may relax the required statistical assumptions. These three contributions facilitate applying the methodology in a variety of settings. An empirical application on attitudes toward LGBT issues shows the potential of the Crosswise Model. An interactive app, Python and MATLAB codes support broader adoption of the model.
It is commonly held that even where questionnaire response is poor, correlational studies are affected only by loss of degrees of freedom or precision. We show that this supposition is not true. If the decision to respond is correlated with a substantive variable of interest, then regression or analysis of variance methods based upon the questionnaire results may be adversely affected by self-selection bias. Moreover such bias may arise even where response is 100%. The problem in both cases arises where selection information is passed to the score indirectly via the disturbance or individual effects, rather than entirely via the observable explanatory variables. We suggest tests for the ensuing self-selection bias and possible ways of handling the ensuing problems of inference.
I develop a survey method for estimating social influence over individual political expression, by combining the content-richness of document scaling with the flexibility of survey research. I introduce the “What Would You Say?” question, which measures self-reported usage of political catchphrases in a hypothetical social context, which I manipulate in a between-subjects experiment. Using Wordsticks, an ordinal item response theory model inspired by Wordfish, I estimate each respondent’s lexical ideology and outspokenness, scaling their political lexicon in a two-dimensional space. I then identify self-censorship and preference falsification as causal effects of social context on respondents’ outspokenness and lexical ideology, respectively. This improves upon existing survey measures of political expression: it avoids conflating expressive behavior with populist attitudes, it defines preference falsification in terms of code-switching, and it moves beyond trait measures of self-censorship, to characterize relative shifts in the content of expression between different contexts. I validate the method and present experiments demonstrating its application to contemporary concerns about self-censorship and polarization, and I conclude by discussing its interpretation and future uses.
A key objective for upcoming surveys, and when re-analysing archival data, is the identification of variable stellar sources. However, the selection of these sources is often complicated by the unavailability of light curve data. Utilising a self-organising map (SOM), we demonstrate the selection of diverse variable source types from a catalogue of variable and non-variable SDSS Stripe 82 sources whilst employing only the median $u-g$, $g-r$, $r-i$, and $i-z$ photometric colours for each source as input, without using source magnitudes. This includes the separation of main sequence variable stars that are otherwise degenerate with non-variable sources ($u-g$,$g-r$) and ($r-i$,$i-z$) colour-spaces. We separate variable sources on the main sequence from all other variable and non-variable sources with a purity of $80.0\%$ and completeness of $25.1\%$, figures which can be modified depending on the application. We also explore the varying ability of the same method to simultaneously select other types of variable sources from the heterogeneous sample, including variable quasars and RR-Lyrae stars. The demonstrated ability of this method to select variable main sequence stars in colour-space holds promise for application in future survey reduction pipelines and for the analysis of archival data, where light curves may not be available or may be prohibitively expensive to obtain.
This chapter introduces a research design to study the effects of community policing. The chapter introduces the Metaketa model of multi-site trials, which are used to answer questions relevant to policy using coordinated experiments in which the same intervention is randomly assigned to units in multiple contexts and the same outcomes are measured to estimate effects. In specific, the chapter introduces how the six countries were selected for study and describes their characteristics in terms of crime and policing and then how the interventions were selected and harmonized across the settings and how they compare to community policing policies in the world. The remainder of the chapter details the experimental design, from how police beats and units are sampled, how community policing intervention was randomly assigned, how outcomes were measured and harmonized, how effects were estimated for each site and then averaging across sites, and how we planned to address threats to inference.
Once the theory is specified and an operationalization has been chosen for the nodes and links, the next step is to acquire the data. This chapter goes deep into issues that arise when designing surveys to collect data. Although this is not the only method of data collection, it is one that illuminates issues that pertain to all others. This chapter covers the practical question of how to use surveys to elicit network information. The advice leans heavily on a well-formulated theory.
The importance of simple descriptive data was recognised by William Farr, whom we mentioned briefly in Chapter 1 for his seminal work using the newly established vital statistics register of England in the nineteenth century. As we discussed in Chapter 1, this descriptive epidemiology, concerned as it is with ‘person, place and time’, attempts to answer the questions ‘Who?’, ‘What?’, ‘Where?’ and ‘When?’. This can include anything from a description of disease in a single person (a case report) or a special survey conducted to measure the prevalence of a particular health issue in a specific population, to reports from national surveys and data collection systems showing how rates of disease or other health-related factors vary in different geographical areas or over time (time trends). In this chapter we look in more detail at some of the most common types of descriptive data and where they come from. However, before embarking on a data hunt, we first need to decide exactly what it is we want to know, and this can pose a challenge. To make good use of the most relevant descriptive data, it is critical to formulate our question as precisely as possible.
Taking a simplified approach to statistics, this textbook teaches students the skills required to conduct and understand quantitative research. It provides basic mathematical instruction without compromising on analytical rigor, covering the essentials of research design; descriptive statistics; data visualization; and statistical tests including t-tests, chi-squares, ANOVAs, Wilcoxon tests, OLS regression, and logistic regression. Step-by-step instructions with screenshots are used to help students master the use of the freely accessible software R Commander. Ancillary resources include a solutions manual and figure files for instructors, and datasets and further guidance on using STATA and SPSS for students. Packed with examples and drawing on real-world data, this is an invaluable textbook for both undergraduate and graduate students in public administration and political science.
This paper distinguishes news about short-lived events from news about changes in longer term prospects using surveys of expectations. Employing a multivariate GARCH-in-Mean model for the US, the paper illustrates how the different types of news influence business cycle dynamics. The influence of transitory output shocks can be relatively large on impact but gradually diminishes over two to three years. Permanent shocks drive the business cycle, generating immediate stock price reactions and gradually building output effects, although they have more immediate output effects during recessions through the uncertainties they create. Markedly different macroeconomic dynamics are found if these explicitly identified types of news or uncertainty feedbacks are omitted from the analysis.
Chapter 2 covers the period in the late 1960s and early 1970s when the Brazilian military government planned and began building its big dams. It argues that political pressures encouraged the military regime to build dams with giant reservoirs and to do so quickly and without regard to their social and environmental footprints. The dictatorship looked to hydropower projects as a means of powering industrial and economic growth that would legitimize military rule, and it wasted no time in starting construction because it takes a long time to build big dams, often the better part of a decade, and sometimes longer. The 1973 oil crisis added urgency, raising the price of imported petroleum and pushing the government to invest in alternative sources of energy. The crisis encouraged the military regime to double down on the big dams already under construction and to plan a host of new ones. Political pressures also made their way into debates about specific dam sites. The most prominent case was the binational Itaipu Dam (on the Brazilian-Paraguayan Border), where the military government had to weigh geopolitical considerations alongside other criteria. The result of all these political pressures combined was a firm commitment to building large reservoirs in environmentally sensitive areas without public debate and without completing thorough environmental impact studies.
In working with network data, data acquisition is often the most basic yet the most important and challenging step. The availability of data and norms around data vary drastically across different areas and types of research. A team of biologists may spend more than a decade running assays to gather a cells interactome; another team of biologists may only analyze publicly available data. A social scientist may spend years conducting surveys of underrepresented groups. A computational social scientist may examine the entire network of Facebook. An economist may comb through large financial documents to gather tables of data on stakes in corporate holdings. In this chapter, we move one step along the network study life-cycle. Key to data gathering is good record-keeping and data provenance. Good data gathering sets us up for future success—otherwise, garbage in, garbage out—making it critical to ensure the best quality and most appropriate data is used to power your investigation.
This chapter tests two ways of overcoming uncertainty about relationality – having potential collaborators directly communicate how they will relate to each other, and using third parties such as matchmakers and boundary spanners. Both are useful for creating valuable new collaborative relationships, especially between people who begin as strangers. In addition, this chapter also presents evidence showing the impact of new collaborative relationships on strategic decision-making. Data in this chapter come from a variety of national surveys, field experiments, and case comparisons.
For the vast majority of us salary is one of the key reasons we work. With bills to pay and mouths to feed we want to know we are being fairly compensated for the jobs we do. One way to check this is to use the BIALL Annual Salary Survey which considers not only salaries but also working conditions and benefits, offering a comparison to similar organisations and roles. In this article Julie Christmas, Claire Mazer and the team from CB Resourcing highlight some of the findings from the 2023 survey and consider the reasons behind these.
Survey experiments often yield intention-to-treat effects that are either statistically and/or practically “non-significant.” There has been a commendable shift toward publishing such results, either to avoid the “file drawer problem” and/or to encourage studies that conclude in favor of the null hypothesis. But how can researchers more confidently adjudicate between true, versus erroneous, nonsignificant results? Guidance on this critically important question has yet to be synthesized into a single, comprehensive text. The present essay therefore highlights seven “alternative explanations” that can lead to (erroneous) nonsignificant findings. It details how researchers can more rigorously anticipate and investigate these alternative explanations in the design and analysis stages of their studies, and also offers recommendations for subsequent studies. Researchers are thus provided with a set of strategies for better designing their experiments, and more thoroughly investigating their survey-experimental data, before concluding that a given result is indicative of “no significant effect.”
Chapter 6 presents an array of techniques for assessing motivation including self-reports, questionnaires, rating scales, checklists, surveys, interviews, and a diagnostic protocol. In addition to these assessments, Appendices 6A, 6B, and 6C – designed for teachers, students, and corporate folk – contain the Reisman Diagnostic Motivation Assessment (RDMA) items that emerged from and are categorized by the motivation theorists presented in Chapter 3. Appendix 6D includes the Reisman Diagnostic Creativity Assessment (RDCA) interpretation and Appendix 6E provides an alternative RDMA interpretation. This chapter also addresses why motivation in education is important, students and motivation, teachers and motivation, corporate employer and employee motivation, self-esteem and motivation, and motivation and creativity.