Type: | Package |
Title: | Smoothing Methods for Nonparametric Regression and Density Estimation |
Version: | 2.2-6.0 |
Date: | 2024-02-17 |
Author: | Adrian Bowman and Adelchi Azzalini. Ported to R by B. D. Ripley <ripley@stats.ox.ac.uk> up to version 2.0, version 2.1 by Adrian Bowman and Adelchi Azzalini, version 2.2 by Adrian Bowman. |
Maintainer: | Adrian Bowman <adrian.bowman@glasgow.ac.uk> |
Depends: | R (≥ 3.1.0) |
Suggests: | rgl, misc3d, interp, gam, tkrplot, rpanel (≥ 1.1-4), tcltk |
Description: | This is software linked to the book 'Applied Smoothing Techniques for Data Analysis - The Kernel Approach with S-Plus Illustrations' Oxford University Press. |
License: | GPL-2 | GPL-3 [expanded from: GPL (≥ 2)] |
LazyData: | TRUE |
Encoding: | UTF-8 |
RoxygenNote: | 7.2.1 |
NeedsCompilation: | yes |
Packaged: | 2024-02-17 15:23:19 UTC; adrianbowman |
Repository: | CRAN |
Date/Publication: | 2024-02-17 17:40:02 UTC |
These data record six characteristics of aircraft designs which appeared during the twentieth century
Description
These data record six characteristics of aircraft designs which appeared during the twentieth century.
The variables are:
Yr | year of first manufacture |
Period | a code to indicate one of three broad time periods |
Power | total engine power (kW) |
Span | wing span (m) |
Length | length (m) |
Weight | maximum take-off weight (kg) |
Speed | maximum speed (km/h) |
Range | range (km) |
Source: The data were collected by P. Saviotti and are described in detail in Saviotti (1996), "Technological Evolution, Variety and Economy", Edward Elgar: Cheltenham.
These data list the first two principal component scores from the aircraft data, which record six characteristics of aircraft designs throughout the twentieth century
Description
These data list the first two principal component scores from the aircraft data, which record six characteristics of aircraft designs throughout the twentieth century.
The variables are:
Comp.1: | first principal component score |
Comp.2: | second principal component score |
Yr: | year of first manufacture |
Period: | a code to indicate one of three broad time periods |
The data were collected by P. Saviotti and are described in detail in Saviotti (1996), "Technological Evolution, Variety and Economy", Edward Elgar: Cheltenham.
Construct frequency table from raw data in 1, 2 or 3 dimensions.
Description
Given a vector, or a matrix with 1, 2 or 3 columns, this function constructs a frequency table
associated with appropriate intervals covering the range of x
.
Usage
binning(x, y, breaks, nbins)
Arguments
x |
a vector or a matrix with either one, two or three columns, containing the original data. |
y |
a vector of data, for example response data, associated with the data in |
breaks |
either a vector or a matrix with two columns (depending on the dimension of |
nbins |
the number of intervals on each axis. If |
Details
This function is called automatically (under the default settings)
by some of the functions of the sm
library when the sample size is
large, to allow handling of datasets of essentially unlimited size.
Specifically, it is used by sm.density
, sm.regression
, sm.ancova
,
sm.binomial
and sm.poisson
.
Value
In the vector case, a list is returned containing the following elements:
a vector x
of the midpoints of the bins excluding those with 0 frequecies,
its associated matrix x.freq
of frequencies, the co-ordinates of the
midpoints
, the division points, and the complete vector of observed
frequencies freq.table
(including the 0 frequencies), and the vector
breaks
of division points.
In the matrix case, the returned value is a list with the following
elements: a two-dimensional matrix x
with the coordinates of the
midpoints of the two-dimensional bins excluding those with 0 frequencies,
its associated matrix x.freq
of frequencies, the coordinates of the
midpoints
, the matrix breaks
of division points, and the observed
frequencies freq.table
in full tabular form.
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
See Also
sm
, sm.density
, sm.regression
, sm.binomial
, sm.poisson
, cut
, table
Examples
# example of 1-d use
x <- rnorm(1000)
xb <- binning(x)
xb <- binning(x, breaks=seq(-4,4,by=0.5))
# example of 2-d use
x <- rnorm(1000)
y <- 2*x + 0.5*rnorm(1000)
x <- cbind(x, y)
xb<- binning(x, nbins=12)
Low birthweight in babies
Description
The data refer to a study on low birthweight in babies, defined as "less than 2500 grams", and related factors.
The variables are:
Smoke | indicator of whether the mother is a smoker |
Lwt | last menstrual weight of the mother |
Low | indicator variable of low weight |
Hosmer & Lemeshow (1989). Applied Logistic Regression. Wiley, NY. The original source contains additional variables; see Appendix 1 of Hosmer & Lemeshow for a full list of the data, pp.29-30 and p.92 for additional information.
Flaws in cloth
Description
These data refer to the length and the observed number of flaws in rolls of cloth.
The variables are:
Length | length of each roll (m) |
Flaws | number of flaws detected |
Source: Bissell (1972). A negative binomial model with varying element sizes. Biometrika 59, 435-41.
Yield-density relationship for Brown Imperial Spanish onions
Description
These data were collected in a study of the relationship between the yield of Brown Imperial Spanish onion plants and the density of planting.
The variables are:
Density | density of planting (plants/m^2) |
Yield | yield (g/plant) |
Locality | a code to indicate Mount Gambier (1) or Uraidla (2) |
The data were collected by I.S.Rogers (South Australian Dept. of Agriculture & Fisheries). They are listed in Ratkowsky (1983), Nonlinear Regression Modeling. Dekker, New York.
Coastline of the UK and Ireland
Description
These data provide the co-ordinates of a set of points lying on the coastlines of the UK and Ireland.
The variables are:
britlong | longitude |
brotlat | latitude |
The relationship between plasma citrate and carbohydrate metabolites
Description
These data were collected in an experiment to study the relationship between possible daily rhythms of plasma citrate and daily rhythms of carbohydrate metabolites during feeding with a citrate-poor diet. During the experiment, plasma citrate concentrations were determined for each of 10 subjects at 14 successive time points during the day. The measurements covered the period 8a.m. to 9p.m. at hourly intervals. Meals were given at 8a.m., noon and 5p.m.
The variables are denoted by C08
, ..., C21
and refer to plasma citrate measurements at the indicated hours.
Anderson,A.H., Jensen,E.B. & Schou,G.(1981). Two-way analysis of variance with correlated errors. Int.Stat.Rev. 49,153-67.
The data were taken from a report by T.T.Nielsen, N.S.Sorensen and E.B.Jensen.
Coal ash in mining samples
Description
These data record the percentage of coal ash found in mining samples originally reported by Gomez and Hazen (1970) and subsequently used by Cressie (1993).
The variables are:
East | a code for east-west direction |
North | a code for north-south direction |
Percent | the percentage of coalash |
Cressie, N. (1993). Statistics for Spatial Data, revised edition. New York: Wiley.
Gomez, M. and Hazen, K. (1970). Evaluating sulfur and ash distribution in coal seams by statistical response surface regression analysis. Report RI 7377, U.S. Bureau of Mines, Washington, D. C.
Coronary sinus potassium in dogs
Description
Measurements of coronary sinus potassium (mil equivalent per litre) were made at (1,3,5,7,9,11,13) minutes after coronary occlusion in a number of different dogs. There are four treatment groups (group 1 is the control). The paper by Grizzle and Allen provides a full description of the treatments. A few subjects develop ventricular fibrillation (see paper for details).
The variables are:
Group | a treatment group indicator |
P1, P3, P5, P7, P9, P11, P13 | measurements at indicated times |
J.E.Grizzle & D.M.Allen (1969). Analysis of growth and dose response curves. Biometrics vol.25, p.357-381
Ovarian follicle counts
Description
These data record the number of ovarian follicles, on a log scale, counted from sectioned ovaries of women of various ages.
The variables are:
Age | age of the woman |
Count | follicle count |
Source | an indicator of the source of the data |
The data were reported by Block (1952; 1953), Richardson et al. (1987) and A Gougeon. They are analysed by Faddy & Gosden (1996) and Faddy & Jones (1997).
Duration and the time between eruptions for the Old Faithful Geyser
Description
These data document the duration of eruptions, and the time between eruptions, for the Old Faithful Geyser in Yellowstone National Park.
The variables are:
Waiting | the waiting time before each eruption (minutes) |
Next.waiting | the waiting time following each eruption (minutes) |
Duration | the length of an eruption ( minutes) |
The data were collected by by the Park Geologist, R.A.Hutchinson. An earlier set of data is reported in Weisberg (1990), Applied Linear Regression, Wiley, New York. The later set, used here, was reported by Azzalini & Bowman (1990), "A look at some data on the Old Faithful Geyser", Applied Statistics 39, 357-65.
Old Faithful Geyser Data
Description
A version of the eruptions data from the “Old Faithful” geyser in Yellowstone National Park, Wyoming. This version comes from Azzalini and Bowman (1990) and is of continuous measurement from 1st August to 15th August 1985.
Some nocturnal duration measurements were coded as 2, 3 or 4 minutes, having originally been described as ‘short’, ‘medium’ or ‘long’.
Usage
data(geyser)
Format
A data frame with 299 observations on 2 variables.
duration | numeric | Eruption time, in minutes |
waiting | numeric | Waiting time before the eruption, in minutes |
Note that in versions of the sm
package before 2.2-5 the waiting
variable was incorrectly described as ‘waiting time to next eruption’.
References
Azzalini, A. and Bowman, A. W. (1990) A look at some data on the Old Faithful geyser. Applied Statistics 39, 357–365.
See Also
Selection of the smoothing parameter
Description
This function selects a smoothing parameter for density estimation in one or two dimensions and for nonparametric regression with one or two covariates. Several methods of selection are available.
Usage
h.select(x, y = NA, weights = NA, group = NA, ...)
Arguments
x |
a vector, or two-column matrix. |
y |
a vector of responses, in regression case. |
weights |
a vector of integers representing frequencies of individual
observations. Use of this parameter is incompatible with
|
group |
a vector of groups indicators (numeric or character values) or a factor |
... |
other optional parameters are passed to the |
Details
see the two references below for discussion of the methods of smoothing parameter selection.
If the sample size is large, binning will be employed. In the
case of method = "cv"
the answer will therefore be
different from that obtained through the function hcv
where binning is not used.
When the group
argument is set, the chosen method of
smoothing parameter selection is applied to each group and the
value returned is the geometric mean of these. This is intended
for use in sm.density.compare
and
sm.ancova
, where
the same smoothing parameter is used for all groups so that
the principal bias terms cancel when the estimates are compared.
Value
the value of the selected smoothing parameter.
Side Effects
none
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
Hurvich, C.M., Simonoff, J.S. and Tsai, C.-L. (1998). Smoothing parameter selection in nonparametric regression using an improved Akaike information criterion. J. R. Statistic. Soc., Series B, 60, 271-293.
See Also
Examples
x <- rnorm(50)
h.select(x)
h.select(x, method = "sj")
x <- matrix(rnorm(100), ncol = 2)
h.select(x)
sm.density(x, method = "cv")
x <- rnorm(50)
y <- x^2 + rnorm(50)
h.select(x, y)
sm.regression(x, y, method = "aicc")
x <- matrix(rnorm(100), ncol = 2)
y <- x[,1]^2 + x[,2]^2 + rnorm(50)
h.select(x, y, method = "cv", structure.2d = "common")
sm.regression(x, y, df = 8)
Cross-validatory choice of smoothing parameter
Description
This function uses the technique of cross-validation to select a smoothing parameter suitable for constructing a density estimate or nonparametric regression curve in one or two dimensions.
Usage
hcv(x, y = NA, hstart = NA, hend = NA, ...)
Arguments
x |
a vector, or two-column matrix of data. If |
y |
a vector of response values for nonparametric regression. |
hstart |
the smallest value of the grid points to be used in an initial grid search for the value of the smoothing parameter. |
hend |
the largest value of the grid points to be used in an initial grid search for the value of the smoothing parameter. |
... |
other optional parameters are passed to the |
Details
See Sections 2.4 and 4.5 of the reference below.
The two-dimensional case uses a smoothing parameter derived from a single value, scaled by the standard deviation of each component.
This function does not employ a sophisticated algorithm and some
adjustment of the search parameters may be required for different sets
of data. An initial estimate of the value of h which minimises the
cross-validatory criterion is located from a grid search using values
which are equally spaced on a log scale between hstart
and
hend
. A quadratic approximation is then used to refine this
initial estimate.
Value
the value of the smoothing parameter which minimises the cross-validation criterion over the selected grid.
Side Effects
If the minimising value is located at the end of the grid of search positions,
or if some values of the cross-validatory criterion cannot be evaluated,
then a warning message is printed. In these circumstances altering the
values of hstart
and hend
may improve performance.
Note
As from version 2.1 of the package, a similar effect can be
obtained with the new function h.select
, via
h.select(x, method="cv")
. Users are encouraged to adopt
this route, since hcv
might be not accessible directly
in future releases of the package. When the
sample size is large hcv
uses the raw data while
h.select(x, method="cv")
uses binning. The latter is
likely to produce a more stable choice for h
.
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
See Also
Examples
# Density estimation
x <- rnorm(50)
par(mfrow=c(1,2))
h.cv <- hcv(x, display="lines", ngrid=32)
sm.density(x, h=hcv(x))
par(mfrow=c(1,1))
# Nonparametric regression
x <- seq(0, 1, length = 50)
y <- rnorm(50, sin(2 * pi * x), 0.2)
par(mfrow=c(1,2))
h.cv <- hcv(x, y, display="lines", ngrid=32)
sm.regression(x, y, h=hcv(x, y))
par(mfrow=c(1,1))
Normal optimal choice of smoothing parameter in density estimation
Description
This functions evaluates the smoothing parameter which is asymptotically optimal for estimating a density function when the underlying distribution is Normal. Data in one, two or three dimensions can be handled.
Usage
hnorm(x, weights)
Arguments
x |
a vector, or matrix with two or three columns, containing the data. |
weights |
an optional vector of integer values which allows the kernel functions over the observations to take different weights when they are averaged to produce a density estimate. This is useful, in particular, for censored data and to construct an estimate from binned data. |
Details
See Section 2.4.2 of the reference below.
Value
the value of the asymptotically optimal smoothing parameter for Normal case.
Note
As from version 2.1 of the package, a similar effect can be
obtained with the new function h.select
, via h.select(x,
method="normal", weights=weights)
or simply h.select(x)
.
Users are encouraged to adopt this route, since hnorm
might be
not accessible directly in future releases of the package.
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
See Also
Examples
x <- rnorm(50)
hnorm(x)
Sheather-Jones choice of smoothing parameter for density estimation
Description
This function uses the Sheather-Jones plug-in method of selecting a smoothing parameter which is suitable for constructing a density estimate in the one-dimensional case.
Usage
hsj(x)
Arguments
x |
a vector of data. |
Details
See Section 2.4.4 of the reference below.
Value
the value of the smoothing parameter located by the Sheather-Jones method.
Note
As from version 2.1 of the package, a similar effect can be
obtained with the new function h.select
, via
h.select(x, method="sj")
. Users are encouraged to adopt
this route, since hsj
might be not accessible directly
in future releases of the package.
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
See Also
Examples
x <- rnorm(50)
hsj(x)
Spatial positions of cases of laryngeal cancer
Description
These data record the spatial positions of cases of laryngeal cancer in the North-West of England between 1974 and 1983, together with the positions of a number of lung cancer patients who were used as controls. The data have been adjusted to preserve anonymity.
The variables are:
Easting | a west-east grid reference |
Northing | a north-south grid reference |
Cancer | an indicator of laryngeal (1) or lung (2) cancer |
Source: Bailey & Gatrell (1995). Interactive Spatial Data Analysis. Longman Scientific and Technical, Harlow. A more extensive set of data is analysed in Kelsall & Diggle, kernel estimation of relative risk, Bernoulli 1, 3-16.
The abundance of mackerel eggs
Description
These data record the abundance of mackerel eggs off the coast of north-western Europe, from a multi-country survey in 1992.
The variables are:
Density | egg density |
mack.lat | latitude of sampling position |
mack.long | longitude of sampling position |
mack.depth | bottom depth |
Temperature | surface temperature |
Salinity | salinity |
Background to the survey and the data are provided by Watson et al. (1992), Priede and Watson (1993) and Priede et al (1995). Borchers et al (1997) describe an analysis of the data.
Magnetic remanence
Description
These data record measurements of magnetic remanence in specimens of Precambrian volcanics.
The variables are:
maglong | directional component on a longitude scale |
maglat | directional component on a latitude scale |
Schmidt & Embleton (1985) J.Geophys.Res. 90 (B4), 2967-2984.
The data are also listed in Fisher, Lewis & Embleton (1987), Statistical Analysis of Spherical Data, Cambridge University Press, Cambridge, dataset B6.
Mildew control
Description
The data refer to study of mildew control sponsored by Bainbridge, Jenkyn and Dyke at Rothamsted Experimental Station. There were four treatments, one of which was a control. There were 36 adjacent plots, with an extra plot at each end. Nine blocks were created by grouping the plots in fours.
The variables are:
t1, t2, t3 | indicators of the four treatment groups |
p1, ..., p8 | indicators of the nine blocks |
Yield | tons of grain per hectare |
Draper & Guttman (1980). Incorporating overlap effects from neighbouring units into response surface models. Applied Statistics 29, 128-134.
Heavy metals in mosses in Galicia.
Description
Mosses are used as a means of measuring levels of heavy metal concentrations in the atmosphere, since most of the nutrient uptake of the mosses is from the air. This technique for large-scale monitoring of long-range transport processes has been used in Galicia, in North-West Spain, over the last decade, as described by Fernandez et al. (2005). In 2006, in both March and September, measurements of different metals were collected at 148 points lying almost in a regular grid over the region with 15 km spacing in north-south and east-west directions. According to the ecologists' expertise, the period between the two samples, passing from a humid to a dry season, is enough time to guarantee the independence of the observed processes.
The dataset consists of a list with six components
loc.m | a two-column matrix containing grid locations of the March monitoring sites |
loc.s | a two-column matrix containing grid locations of the September monitoring sites |
Co.m | cobalt concentration (log scale) in March |
Co.s | cobalt concentration (log scale) in September |
Hg.m | mercury concentration (log scale) in March |
Hg.s | mercury concentration (log scale) in September |
Source: The data were kindly made available by the Ecotoxicology and Vegetal Ecophysiology research group in the University of Santiago de Compostela.
References
Fernandez, J., Real, C., Couto, J., Aboal, J., Carballeira, A. (2005). The effect of sampling design on extensive biomonitoring surveys of air pollution. Science of the Total Environment, 337, 11-21.
Examples
## Not run:
# Comparison of Co in March and September
with(mosses, {
nbins <- 12
vgm.m <- sm.variogram(loc.m, Co.m, nbins = nbins, original.scale = TRUE,
ylim = c(0, 1.5))
vgm.s <- sm.variogram(loc.s, Co.s, nbins = nbins, original.scale = TRUE,
add = TRUE, col.points = "blue")
trns <- function(x) (x / 0.977741)^4
del <- 1000
plot(vgm.m$distance.mean, trns(vgm.m$sqrtdiff.mean), type = "b",
ylim = c(0, 1.5), xlab = "Distance", ylab = "Semi-variogram")
points(vgm.s$distance.mean - del, trns(vgm.s$sqrtdiff.mean), type = "b",
col = "blue", pch = 2, lty = 2)
plot(vgm.m$distance.mean, trns(vgm.m$sqrtdiff.mean), type = "b",
ylim = c(0, 1.5), xlab = "Distance", ylab = "Semi-variogram")
points(vgm.s$distance.mean - del, trns(vgm.s$sqrtdiff.mean), type = "b",
col = "blue", pch = 2, lty = 2)
segments(vgm.m$distance.mean, trns(vgm.m$sqrtdiff.mean - 2 * vgm.m$se),
vgm.m$distance.mean, trns(vgm.m$sqrtdiff.mean + 2 * vgm.m$se))
segments(vgm.s$distance.mean - del, trns(vgm.s$sqrtdiff.mean - 2 * vgm.s$se),
vgm.s$distance.mean - del, trns(vgm.s$sqrtdiff.mean + 2 * vgm.s$se),
col = "blue", lty = 2)
mn <- (vgm.m$sqrtdiff.mean + vgm.s$sqrtdiff.mean) / 2
se <- sqrt(vgm.m$se^2 + vgm.s$se^2)
plot(vgm.m$distance.mean, trns(vgm.m$sqrtdiff.mean), type = "n",
ylim = c(0, 1.5), xlab = "Distance", ylab = "Semi-variogram")
polygon(c(vgm.m$distance.mean, rev(vgm.m$distance.mean)),
c(trns(mn - se), rev(trns(mn + se))),
border = NA, col = "lightblue")
points(vgm.m$distance.mean, trns(vgm.m$sqrtdiff.mean))
points(vgm.s$distance.mean, trns(vgm.s$sqrtdiff.mean), col = "blue", pch = 2)
vgm1 <- sm.variogram(loc.m, Co.m, nbins = nbins, varmat = TRUE,
display = "none")
vgm2 <- sm.variogram(loc.s, Co.s, nbins = nbins, varmat = TRUE,
display = "none")
nbin <- length(vgm1$distance.mean)
vdiff <- vgm1$sqrtdiff.mean - vgm2$sqrtdiff.mean
tstat <- c(vdiff %*% solve(vgm1$V + vgm2$V) %*% vdiff)
pval <- 1 - pchisq(tstat, nbin)
print(pval)
})
# Assessing isotropy for Hg in March
with(mosses, {
sm.variogram(loc.m, Hg.m, model = "isotropic")
})
# Assessing stationarity for Hg in September
with(mosses, {
vgm.sty <- sm.variogram(loc.s, Hg.s, model = "stationary")
i <- 1
image(vgm.sty$eval.points[[1]], vgm.sty$eval.points[[2]], vgm.sty$estimate[ , , i],
col = topo.colors(20))
contour(vgm.sty$eval.points[[1]], vgm.sty$eval.points[[2]], vgm.sty$sdiff[ , , i],
col = "red", add = TRUE)
})
## End(Not run)
Rat skeletal muscles
Description
The data refer to the counts of fibres of different types in groups of fibres taken from rat skeletal muscles.
The variables are:
row.labels | indicators of the four treatment groups |
TypeI.R | indicators of the nine blocks |
TypeI.P | tons of grain per hectare |
TypeI.B | tons of grain per hectare |
TypeII | tons of grain per hectare |
Hand, Daly, Lunn, McConway and Ostrowski (1994). A handbook of Small Data Sets. Chapman & Hall: London. The data were collected by M. Khan and M. Khan.
Water level of the River Nile
Description
These data record historical data on the water level of the River Nile.
The variables are:
Volume | Annual volume of the Nile River (discharge at Aswan, 10^8 m^3) |
Year | 1871-1970 |
Cobb, G. (1978). The problem of the Nile: conditional solution to a change-point problem. Biometrika 65, 243-251.
Integrated squared error between a density estimate and a Normal density
Description
This function evaluates the integrated squared error between a density
estimate constructed from a standardised version of the univariate data
y
and a standard normal density function.
Usage
nise(y, ...)
Arguments
y |
a vector of data. |
... |
further arguments which are to be passed to |
Details
The data y
are first standardised to have sample mean 0 and sample
variance 1. The integrated squared error between a density estimate
constructed from these standardised data and a standard normal distribution
is then evaluated.
See Section 2.5 of the reference below.
Value
the integrated squared error.
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
See Also
Examples
x <- rnorm(100)
nise(x)
mean integrated squared error for density estimation with normal data
Description
This function evaluates the mean integrated squared error of a density estimate which is constructed from data which follow a normal distribution.
Usage
nmise(sd, n, h)
Arguments
sd |
the standard deviation of the normal distribution from which the data arise. |
n |
the sample size of the data. |
h |
the smoothing parameter used to construct the density estimate. |
Details
see Section 2.4 of the reference below.
Value
the mean integrated squared error of the density estimate.
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
See Also
Examples
x <- rnorm(50)
sd <- sqrt(var(x))
n <- length(x)
h <- seq(0.1, 2, length=32)
plot(h, nmise(sd, n, h), type = "l")
nearest neighbour distances from data in one or two dimensions
Description
This function calculates the k
nearest neighbour distance from each
value in x
to the remainder of the data. In two dimensions, Euclidean
distance is used after standardising the data to have unit variance in
each component.
Usage
nnbr(x, k)
Arguments
x |
the vector, or two-column matrix, of data. |
k |
the required order of nearest neighbour. |
Details
see Section 1.7.1 of the reference below.
Value
the vector of nearest neighbour distances.
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
See Also
none.
Examples
x <- rnorm(50)
hw <- nnbr(x, 10)
hw <- hw/exp(mean(log(hw)))
sm.density(x, h.weights=hw)
Pause before continuing execution
Description
If a program produces several plots on the same window, insertion of
pause()
between two plots suspends execution until the <Enter>
key is pressed, to allow inspection of the current plot.
Usage
pause()
Positions of the south pole
Description
These data refer to positions of the south pole determined from the palaeomagnetic study of New Caledonian laterites.
The variables are:
Latitude |
Longitude
|
The data were collected by Falvey and Musgrave. They are listed in Fisher, Lewis & Embleton (1987), Statistical Analysis of Spherical Data, Cambridge University Press, Cambridge, dataset B1.
Making data available as data.frame
Description
This function is no longer available in the sm package. It should be replaced by the use of attach
, if necessary. Each dataset now also has its own help file.
It was a utility function, widely used in the scripts accompanying
the book described below. The function provided access to the dataset identified by name
. For flexibility, the datasets were provided in ASCII form, with the name of each variable listed in the first row of the file. This function reads the files and makes the data available as a data frame.
Usage
provide.data(data, path, options = list())
Arguments
data |
name of the data to be loaded and attached as |
path |
the path where the data and its documentation should be searched for,
The default value is an appropriate sub-directory of the |
options |
A list of options passed to |
Details
the data file is assumed to be called data.dat
and the documentation
file describing the data (if present) is assumed to be called data.doc
.
If the data.frame
is already attached, it is re-attached in the second
position of the search
list.
To set describe=FALSE
for the rest of the current session, use
sm.options(describe=FALSE)
The function can easily be adapted to play a similar role for other packages.
Value
none
Side Effects
messages are printed on the command window, describing progress of
the operation. If describe=TRUE
and a documentation file exists, this
is printed on the command windows or another windows, depending on
the type of platform where the program is executed.
Author(s)
Bowman, A.W. and Azzalini, A.
See Also
data.frame
, attach
, sm
,
sm.options
Examples
provide.data(birth)
Radiocarbon in Irish oak
Description
These data record high precision measurements of radiocarbon on Irish oak, used to construct a calibration curve.
The variables are:
Rc.age | age predicted from the radiocarbon dating process |
Precision | a measure of precision of the dating process |
Cal.age | true calendar age |
Pearson & Qua (1993). High precision 14C measurement of Irish oaks to show the natural 14C variations from AD 1840 - 5000 BC: a correction. Radiocarbon 35, 105-123.
A significance trace for a hypothesis test
Description
This function creates a significance trace for a hypothesis test based on a nonparametric smoothing procedure. The p-value of the test is plotted against a range of smoothing parameters.
Usage
sig.trace(expn, hvec, ...)
Arguments
expn |
an S-Plus expression which should define the hypothesis test to be
performed, with the value of the smoothing parameter |
hvec |
a vector of smoothing parameters for which the test will be performed. |
... |
further arguments which will be passed to |
Details
see Section 5.2 of the reference below.
Only tests involving a univariate smoothing parameter may be used.
Value
a list containing vectors with the smoothing parameters and p-values.
Side Effects
If the largest p-value is greater than 0.05 then a horizontal line at 0.05 will be superimposed on any plot, for reference.
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
See Also
none.
Examples
x <- runif(50, 0, 1)
y <- 5*x^2 + rnorm(50)
sig.trace(sm.regression(x, y, model = "linear", display="none"),
hvec = seq(0.05, 0.3, length = 10))
The sm package: summary information
Description
This package implements nonparametric smoothing methods described in the book of Bowman & Azzalini (1997)
Details
Missing data are allowed; they are simply removed, together with
the associated variates from the same case, if any.
Datasets of arbitrary size can be handled by the current version of
sm.density
, sm.regression
and sm.ancova
, using
binning operations.
Main Features
The functions in the package use kernel methods to construct nonparametric estimates of density functions and regression curves in a variety of settings, and to perform some inferential operations.
Specifically, density estimates can be constructed for 1-, 2- and 3-dimensional data. Nonparametric regression for continuous data can be constructed with one or two covariates, and a variety of tests can be carried out. Several other data types can be handled, including survival data, time series, count and binomial data.
Functions
The main functions are sm.density
and sm.regression
; other
functions intended for direct access by the user are:
h.select
, binning
,
sm.ancova
, sm.autoregression
, sm.binomial
,
sm.binomial.bootstrap
, sm.poisson
, sm.poisson.bootstrap
,
sm.options
, sm.rm
, sm.script
, sm.sphere
,
sm.survival
, sm.ts.pdf
, sm.variogram
, sm.pca
. There are undocumented functions which are called by these.
Scripts
The function sm.script
is used to run a set of examples (called
scripts) presented in the book quoted below. These scripts are
associated with the package but the package can be used independently of
them. The scripts are generally based on the functions of the package
sm
, but a few of them make used of the gam
package.
Requirements
R version >= 3.1.0. The gam
package is used by
some of the scripts launched via sm.script
, but it is not
used by the functions of this package.
Version
This is version 2.2. The most recent version of the package can be obtained from the CRAN archive.
Details
The book by Bowman and Azzalini (1997) provides more detailed and
background information. Algorithmic aspects of the software are
discussed by Bowman & Azzalini (2003). Differences between the first
version of the package, described in the book, and the current one are
summarized in the file history.txt
which is distributed with
the package.
Acknowledgements
Important contributions to prototype versions of functions for some specific techniques included here were made by a succession of students; these include Stuart Young, Eileen Wright, Peter Foster, Angela Diblasi, Mitchum Bock and Adrian Hines. We are grateful for all these interactions. These preliminary version have been subsequently re-written for inclusion in the public release of the package, with the exception of the functions for three-dimensional density estimation, written by Stuart Young. We also thank Luca Scrucca who made useful comments and who has ported the software to XLispStat. We are particularly grateful to Brian Ripley for substantial help in the production of installation files, the creation of MS-Windows versions, initial porting of the software from S-Plus to R and for maintaining the package on CRAN for several years.
Licence
This package and its documentation are usable under the terms of the "GNU General Public License", a copy of which is distributed with the package.
Author(s)
Adrian Bowman (School of Mathematics and Statistics, University of Glasgow, UK) and Adelchi Azzalini (Dept Statistical Sciences, University of Padua, Italy). Please send comments, error reports, etc. to the authors.
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
Bowman, A.W. and Azzalini, A. (2003).
Computational aspects of nonparametric smoothing
with illustrations from the sm
library.
Computational Statistics and Data Analysis, 42, 545–560.
Internal sm functions
Description
Internal sm
functions
Usage
addplot(d, f, theta, phi)
britmap()
change(th, ph)
circle(r)
cv(x, h, ...)
hidplot(invis, theta, phi)
incphi(ph, inc)
inctheta(th, inc)
isInteger(x)
isMatrix(x)
normdens.band(x, h, weights = rep(1, length(x)), options = list())
p.quad.moment(A, Sigma, tobs, ndevs)
smplot.regression(x, y, design.mat, h, r, model, weights, rawdata = list(),
options = list(), ...)
plot2(latitude2, longitude2, theta, phi)
plot2d(d, f, theta, phi)
replace.na(List, comp, value)
sj(x, h)
sm.check.data(x, y = NA, weights = NA, group = NA, ...)
sm.density.1d(x, h = hnorm(x, weights), model = "none", weights,
rawdata = list(x = x), options = list())
sm.density.2d(X, h = hnorm(X, weights), weights = rep(1, length(x)),
rawdata = list(), options = list())
sm.density.3d(x, h = hnorm(x, weights), weights = rep(1, length(x)),
rawdata = list(), options = list())
sm.density.eval.1d(x, h, weights = rep(1, n), options = list())
sm.density.eval.2d(x, y, h, xnew, ynew, eval.type = "points",
weights = rep(1, n), options = list())
sm.density.positive.1d(x, h, weights, options = list())
sm.density.positive.2d(X, h = c(hnorm(log(X[, 1] + delta[1]), weights),
hnorm(log(X[,2] + delta[2]), weights)), eval.type = "points",
weights = rep(1, nrow(X)), options = list())
sm.density.positive.grid(X, h = c(hnorm(log(X[, 1] + delta[1])),
hnorm(log(X[, 2] + delta[2]))), weights=NA, options=list())
sm.glm(x, y, family, h, eval.points, start, offset, options=list())
sm.imageplot(x, y, h, weights, rawdata, options = list())
sm.persplot(x, y, h = hnorm(cbind(x, y), weights), weights, rawdata = list(),
options = list())
sm.regression.1d(x, y, h, design.mat = NA, model = "none",
weights = rep(1, length(x)), rawdata, options = list())
sm.regression.2d(x, y, h, model = "none", weights = rep(1, length(y)), rawdata,
options = list())
sm.regression.eval.1d(x, y, design.mat, h, model = "none",
weights = rep(1, length(x)), rawdata, options = list())
sm.regression.eval.2d (x, y, h, model, eval.points, hull = TRUE,
weights, options = list())
sm.regression.test(x, y, design.mat = NA, h, model = "no.effect",
weights = rep(1,length(y)), rawdata, options = list())
sm.sigweight(x, weights)
sm.sliceplot(x, y, h, weights, rawdata = list(), options = list())
sm.weight(x, eval.points, h, cross = FALSE, weights = rep(1, length(x)), options)
sm.weight2(x, eval.points, h, cross = FALSE, weights = rep(1, nrow(x)),
options = list())
smplot.density(x, h, weights = rep(1, length(x)), rawdata = list(x = x),
options = list())
wmean(x, w)
wvar(x, w)
Details
These are not to be called by the user.
Nonparametric analysis of covariance
Description
This function allows a set of nonparametric regression curves to be compared, both graphically and formally in a hypothesis test. A reference model, used to define the null hypothesis, may be either equality or parallelism. Regression surfaces can also be compared in a test but a graphical display is not produced.
Usage
sm.ancova(x, y, group, h, model = "none", h.alpha = NA, weights=NA,
covar = diag(1/weights), ...)
Arguments
x |
a vector or two-column matrix of covariate values. |
y |
a vector of response values. |
group |
a vector of group indicators. |
h |
the smoothing parameter to be used in the construction of each of the
regression curves. If this is missing the method of smoothing parameter
selection specified by |
model |
a character variable which defines the reference model. The values
|
h.alpha |
the value of the smoothing parameter used when estimating the vertical separations of the curves under the parallelism model. If this is missing, it is set to 2 * r / n, where r denotes the range of the data and n the sample size. |
weights |
case weights; see the documentation of |
covar |
the (estimated) covariance matrix of y. The default value assumes
the data to be independent. Where appropriate, the covariance structure
of |
... |
other optional parameters are passed to the |
Details
see Sections 6.4 and 6.5 of the book by Bowman & Azzalini, and the papers by Young & Bowman listed below. This function is a developed version of code originally written by Stuart Young.
Value
a list containing an estimate of the error standard deviation and, where appropriate, a p-value and reference model. If the parallelism model has been selected then a vector of estimates of the vertical separations of the underlying regression curves is also returned. If a reference band has been requested, the upper and lower boundaries and their common evaluation points are also returned.
Side Effects
a plot on the current graphical device is produced, unless display="none"
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
Young, S.G. and Bowman, A.W. (1995). Nonparametric analysis of covariance. Biometrics 51, 920–931.
Bowman, A.W. and Young, S.G. (1996). Graphical comparison of nonparametric curves. Applied Statistics 45, 83–98.
See Also
sm.regression
, sm.density.compare
, sm.options
Examples
x <- runif(50, 0, 1)
y <- 4*sin(6*x) + rnorm(50)
g <- rbinom(50, 1, 0.5)
sm.ancova(x, y, g, h = 0.15, model = "equal")
Nonparametric estimation of the autoregression function
Description
This function estimates nonparametrically the autoregression function
(conditional mean given the past values) of a time series x
,
assumed to be stationary.
Usage
sm.autoregression(x, h = hnorm(x), d = 1, maxlag = d, lags,
se = FALSE, ask = TRUE)
Arguments
x |
vector containing the time series values. |
h |
the bandwidth used for kernel smoothing. |
d |
number of past observations used for conditioning; it must be 1 (default value) or 2. |
maxlag |
maximum of the lagged values to be considered (default value is |
lags |
if |
se |
if |
ask |
if |
Details
see Section 7.3 of the reference below.
Value
a list with the outcome of the final estimation (corresponding to
the last value or pairs of values of lags), as returned by sm.regression
.
Side Effects
graphical output is produced on the current device.
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
See Also
Examples
sm.autoregression(log(lynx), maxlag=3, se=TRUE)
sm.autoregression(log(lynx), lags=cbind(2:3,4:5))
Nonparametric logistic regression
Description
This function estimates the regression curve using the local likelihood approach for a vector of binomial observations and an associated vector of covariate values.
Usage
sm.binomial(x, y, N = rep(1, length(y)), h, ...)
Arguments
x |
vector of the covariate values |
y |
vector of the response values; they must be
nonnegative integers not larger than those of |
h |
the smoothing parameter; it must be positive. |
N |
a vector containing the binomial denominators. If missing, it is assumed to contain all 1's. |
... |
other optional parameters are passed to the |
Details
see Sections 3.4 and 5.4 of the reference below.
Value
A list containing vectors with the evaluation points, the corresponding probability estimates, the linear predictors, the upper and lower points of the variability bands (on the probability scale) and the standard errors on the linear predictor scale.
Side Effects
graphical output will be produced, depending on the value of the
display
parameter.
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
See Also
sm.binomial.bootstrap
, sm.poisson
,
sm.options
, glm
, binning
Examples
## Not run:
# the next example assumes that all binomial denominators are 1's
sm.binomial(dose, failure, h=0.5)
# in the next example, (some of) the dose levels are replicated
sm.binomial(dose, failure, n.trials, h=0.5)
## End(Not run)
with(birth, {
sm.binomial(Lwt[Smoke=="S"], Low[Smoke=="S"], h=20,
xlab='mother weight[Smoke=="S"]')
x<- seq(0,1,length=30)
y<- rbinom(30,10,prob=2*sin(x)/(1+x))
sm.binomial(x,y,N=rep(10,30), h=0.25)
})
Bootstrap goodness-of-fit test for a logistic regression model.
Description
This function is associated with sm.binomial
for the underlying fitting
procedure. It performs a Pseudo-Likelihood Ratio Test for the
goodness-of-fit of a standard parametric logistic regression of specified
degree
in the covariate x
.
Usage
sm.binomial.bootstrap(x, y, N = rep(1, length(x)), h, degree = 1,
fixed.disp=FALSE, ...)
Arguments
x |
vector of the covariate values |
y |
vector of the response values; they must be nonnegative integers. |
h |
the smoothing parameter; it must be positive. |
N |
a vector containing the binomial denominators. If missing, it is assumed to contain all 1's. |
degree |
specifies the degree of the fitted polynomial in |
fixed.disp |
if |
... |
additional parameters passed to |
Details
see Section 5.4 of the reference below.
Value
a list containing the observed value of the Pseudo-Likelihood Ratio Test statistic, its observed p-value as estimated via the bootstrap method, and the vector of estimated dispersion parameters when this value is not forced to be 1.
Side Effects
Graphical output representing the bootstrap samples is produced on the current graphical device. The estimated dispersion parameter, the value of the test statistic and the observed significance level are printed.
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
See Also
sm.binomial
, sm.poisson.bootstrap
Examples
## Not run: sm.binomial.bootstrap(concentration, dead, N, 0.5, nboot=50)
Nonparametric density estimation in one, two or three dimensions.
Description
This function creates a density estimate from data in one, two or three dimensions. In two dimensions a variety of graphical displays can be selected, and in three dimensions a contour surface can be plotted. A number of other features of the construction of the estimate, and of its display, can be controlled.
If the rpanel
package is available, an interactive panel can be
activated to control various features of the plot.
If the rgl
package is also available, rotatable plots are
available for the two- and three-dimensional cases. (For
three-dimensional data, the misc3d
package is also required.)
Usage
sm.density(x, h, model = "none", weights = NA, group=NA, ...)
Arguments
x |
a vector, or a matrix with two or three columns, containing the data. |
h |
a vector of length one, two or three, defining the smoothing parameter.
A normal kernel function is used and |
model |
This argument applies only with one-dimensional data. Its default value
is |
weights |
a vector of integers representing frequencies of individual observations.
Use of this parameter is incompatible with binning; hence |
group |
a vector of groups indicators (numeric or character values) or a factor. |
... |
other optional parameters are passed to the |
Details
see Chapters 1, 2 and 6 of the reference below.
In the three-dimensional case, the contours of the density estimate are
constructed by the contour3d
function in the misc3d
package of Feng & Tierney.
Value
a list containing the values of the density estimate at the evaluation points,
the smoothing parameter, the smoothing parameter weights and the kernel
weights. For one- and two-dimensional data, the standard error of the estimate
(on the square root scale, where the standard error is approximately constant)
and the upper and lower ends of a variability band are also supplied. Less
information is supplied when the smoothing parameter weights
or kernel weights are not all 1, or when positive
is set to TRUE
.
Side Effects
a plot is produced, unless the option display="none"
is set.
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
See Also
h.select
, hnorm
, hsj
, hcv
,
nise
, nmise
, sm
,
sm.sphere
, sm.regression
,
sm.options
Examples
# A one-dimensional example
y <- rnorm(50)
sm.density(y, model = "Normal")
# sm.density(y, panel = TRUE)
# A two-dimensional example
y <- cbind(rnorm(50), rnorm(50))
sm.density(y, display = "image")
# sm.density(y, panel = TRUE)
# A three-dimensional example
# y <- cbind(rnorm(50), rnorm(50), rnorm(50))
# sm.density(y)
Comparison of univariate density estimates
Description
This function allows a set of univariate density estimates to be compared, both graphically and formally in a permutation test of equality.
Usage
sm.density.compare(x, group, h, model = "none", ...)
Arguments
x |
a vector of data. |
group |
a vector of group labels. If this is not already a factor it will be converted to a factor. |
h |
the smoothing parameter to be used in the construction of each density estimate. Notice that the same smoothing parameter is used for each group. If this value is omitted, the mean of the normal optimal values for the different groups is used. |
model |
the default value is |
... |
other optional parameters are passed to the |
Details
For a general description of the methods involved, see Section 6.2 of the reference below.
The colours and linetypes of the density estimates are set by col
and lty
which can be passed as additional arguments. By default these are set to 1 + 1:ngroup
, where ngroup
is the number of groups represented in the group
variable.
Value
A list is returned containing:
estimate |
a matrix whose rows contain the estimates for each group. |
eval.points |
the grid of common evaluation points for the estimates. |
h |
the common smoothing parameter used in the construction of the estimates. |
levels |
the levels of the group factor. |
col , lty , lwd |
plotting details which can be useful in constructing a legend for the plot; see the examples below. |
When "model"
is set to "equal"
, the output list also contains the p-value (p
) of the test.
When band = TRUE
, and there are only two groups to compare, the output list also contains the upper (upper
) and lower (lower
) edges of the reference band for equality.
Side Effects
a plot on the current graphical device is produced, unless
display="none"
.
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
See Also
sm.density
, sm.ancova
, sm.options
Examples
y <- rnorm(100)
g <- rep(1:2, rep(50,2))
sm.density.compare(y, g)
comp <- sm.density.compare(y, g, model = "equal")
legend("topleft", comp$levels, col = comp$col, lty = comp$lty, lwd = comp$lwd)
The detection of discontinuities in a regression curve or surface.
Description
This function uses a comparison of left and right handed nonparametric regression curves to assess the evidence for the presence of one or more discontinuities in a regression curve or surface. A hypothesis test is carried out, under the assumption that the errors in the data are approximately normally distributed. A graphical indication of the locations where the evidence for a discontinuity is strongest is also available.
Usage
sm.discontinuity(x, y, h, hd, ...)
Arguments
x |
a vector or two-column matrix of covariate values. |
y |
a vector of responses observed at the covariate locations. |
h |
a smoothing parameter to be used in the construction of the nonparametric
regression estimates. A normal kernel
function is used and |
hd |
a smoothing parameter to be used in smoothing the differences of the
left and right sided nonparametric regression estimates. A normal kernel
function is used and |
... |
other optional parameters are passed to the |
Details
The reference below describes the statistical methods used in the function. There are minor differences in some computational details of the implementation.
Currently duplicated rows of x
cause a difficulty in the two covariate case. Duplicated rows should be removed.
Value
a list containing the following items
p |
the p-value for the test of the null hypothesis that no discontinuities are present. |
sigma |
the estimated standard deviation of the errors. |
eval.points |
the evaluation points of the nonparametric
regression estimates. When |
st.diff |
a vector or matrix of standardised differences between the left and right sided estimators at the evaluation points. |
diffmat |
when |
angle |
when |
h |
the principal smoothing parameter. |
hd |
the smoothing parameter used for double-smoothing (see the reference below). |
Side Effects
a plot on the current graphical device is produced, unless the option
display="none"
is set.
References
Bowman, A.W., Pope, A. and Ismail, B. (2006). Detecting discontinuities in nonparametric regression curves and surfaces. Statistics & Computing, 16, 377–390.
See Also
Examples
par(mfrow = c(3, 2))
with(nile, {
sm.discontinuity(Year, Volume, hd = 0)
sm.discontinuity(Year, Volume)
ind <- (Year > 1898)
plot(Year, Volume)
h <- h.select(Year, Volume)
sm.regression(Year[!ind], Volume[!ind], h, add = TRUE)
sm.regression(Year[ ind], Volume[ ind], h, add = TRUE)
hvec <- 1:15
p <- numeric(0)
for (h in hvec) {
result <- sm.discontinuity(Year, Volume, h,
display = "none", verbose = 0)
p <- c(p, result$p)
}
plot(hvec, p, type = "l", ylim = c(0, max(p)), xlab = "h")
lines(range(hvec), c(0.05, 0.05), lty = 2)
})
with(trawl, {
Position <- cbind(Longitude, Latitude)
ind <- (Longitude < 143.8)
# Remove a repeated point which causes difficulty with sm.discontinuity
ind[54] <- FALSE
sm.regression(Position[ind,], Score1[ind], theta = 35, phi = 30)
sm.discontinuity(Position[ind,], Score1[ind], col = "blue")
})
par(mfrow = c(1, 1))
# The following example takes longer to run.
# Alternative values for nside are 32 and 64.
# Alternative values of yjump are 1 and 0.5.
# nside <- 16
# yjump <- 2
# x1 <- seq(0, 1, length = nside)
# x2 <- seq(0, 1, length = nside)
# x <- expand.grid(x1, x2)
# x <- cbind(x1 = x[, 1], x2 = x[, 2])
# y <- rnorm(nside * nside)
# ind <- (sqrt((x[, 1] - 0.5)^2 + (x[, 2] - 0.5)^2) <= 0.25)
# y[ind] <- y[ind] + yjump
# image(x1, x2, matrix(y, ncol = nside))
# sm.discontinuity(x, y, df = 20, add = TRUE)
A test of monotonicity in a regression curve.
Description
This function uses the idea of a ‘critical bandwidth’ to assess the evidence that a regression curve is non-monotonic. A hypothesis test is carried out by bootstrap methods and the empirical p-value is reported. Response variables on a continuous scale or with binomial variation can be handled.
Usage
sm.monotonicity(x, y, N = rep(1, length(y)), h, type = "continuous", ...)
Arguments
x |
a vector of covariate values. |
y |
a vector of responses observed at the covariate locations. |
N |
a vector of sample sizes at the covariate locations, when the responses have a binomial error structure. |
h |
a smoothing parameter to be used in the construction of the nonparametric
regression estimates. A normal kernel
function is used and |
type |
an indicator of whether the response variable is on a |
... |
other optional parameters are passed to the |
Details
The first reference below describes the statistical methods used in the function. The test is an extension of one by Silverman (1986) for density estimation.
Value
a list containing the following items
p |
the p-value for the test of the null hypothesis that the true curve is monotonic. |
hcrit |
the ‘critical’ smoothing parameter. This is the smallest value which, when applied to the observed data, makes the curve monotonic. |
h |
the smoothing parameter used for double-smoothing (see the reference below). |
Side Effects
a plot of the curves generated by the bootstrap procedure is produced,
unless the option display="none"
is set. Those curves which
are non-monotonic, and therefore contribute to the empirical p-value,
are drawn in red.
References
Bowman, A.W., Jones, M.C. and Gijbels, I. (1998). Testing monotonicity of regression. J.Comp.Graph.Stat. 7, 489-500.
See Also
Examples
## Not run:
# Radiocarbon dating data
with(radioc, {
ind <- (Cal.age>5000 & Cal.age<6000)
cal.age <- Cal.age[ind]
rc.age <- Rc.age[ind]
sm.monotonicity(cal.age, rc.age, method = "aicc", nboot = 200)
})
# Hosmer & Lemeshow birth data
with(birth, {
sm.monotonicity(Lwt[Smoke == "N"], Low[Smoke == "N"],
type = "binomial")
})
## End(Not run)
Set or return options of sm library
Description
This function provides a means to control the behaviour of the sm
library such as the colour of the plotted lines, the size of the grid in 2-d estimation, the set of evaluations points, and many others. A list may be given as the only argument, or any number of arguments may be in the name=value
form. If no arguments are specified then the function returns the current settings of all the arguments.
Usage
sm.options(...)
Arguments
... |
A list may be given as the only argument, or any number of arguments may be in the |
- add
-
a logical value which controls whether the estimate is added to the current plot. Its default value is
FALSE
, which creates a new plot. This argument applies only with one-dimensional data or to contour andrgl
created from two-dimensional data. - alpha
-
a value, or a vector of values, lying between 0 and 1, which controls the transparency of the surfaces used to construct an
rgl
plot, when this form if display is requested. In the case of regression with two covariates, a single value foralpha
is used and the default is0.7
. In the case of density estimation with three variables,alpha
should be set to a vector whose length matches the number of contours to be drawn (specified by theprops
argument). In this case the default isseq(1, 0.5, length = length(props))
. - alpha.mesh
-
a parameter, lying between 0 and 1, which controls the transparency of the mesh lines used to construct an
rgl
plot for regression with two covariates. The default value is1
. - band
-
a logical value which controls whether the distance between the nonparametric estimate and a reference model should be indicated as a band (one covariate), or through colour painting of the regression surface (two covariates). This is activated only when a model has been nominated through the
model
parameter. In the case of two covariates, the setting of the argumentcol
has priority overband
. The settingse = TRUE
can also activate this feature. - col
-
the colour used for plotting observed points and estimated curves. Where groups are used, with one covariate,
col
may be set to a vector of colours associated with the groups. In regression with two covariates using anrgl
display,col
may be set to a single colour, or to the values"height"
or"se"
. These latter two setting cause the surface to be painted according to its height or to standard error information; see the information on the parametersse
,se.breaks
andmodel
. In the case of density estimation with three variables,col
can be set to a vector whose length matches the number of contours to be drawn (specified by theprops
argument). - col.band
-
the colour used for the reference band when a
model
is specified in regression with one covariate or in density estimation with a single variable. Default:col.band="cyan"
. - col.mesh
-
the colour used for the ‘wire mesh’ representation plotting observed points in an rgl display for regression with two covariates. This can also be set by the second component of a vector of length two which is set for
col
. Default:col.mesh="black"
. - col.palette
-
the colours used for shading an image plot, or for surface painting in an rgl display, for regression with two covariates. Default:
col.palette=topo.colors(12)
. - col.points
-
the colour used for plotting observed points in a regression with one covariate or an
rgl
display for regression with two covariates. Default:col.points="black"
. - delta
in
sm.density
, a value which will be added to the data before they are log transformed in the procedure to handle positive data. The value ofdelta
is used only whenpositive
takes the valueTRUE
. The default value is the smallest value observed in each dimension. This argument does not apply with three-dimensional data. Default:delta=NA
- describe
-
logical flag which affects the behaviour of
sm.script
andprovide.data
. Ifdescribe=TRUE
(default), a data documentation file is printed. - df
approximate degrees-of-freedom of the smoothing parameter used in
sm.regression
, when a numerical value ofh
is not specified. In this case, the equivalent value ofh
will be computed and included in the list returned on exit fromsm.regression
. Default value is 6 ifx
is a vector and 12 ifx
is a matrix.- diff.ord
in
sm.regression
, an integer defining the degree of differencing to be applied in the estimation process. When this argument is set to 1, the method of Rice, based on the squared differences of pairs of neighbouring observations, is used. When the argument is set to 2 (default), the method of Gasser, Sroka and Jennen-Steinmetz, based on differences between each observation and a linear interpolation from its two neighbours, is used.- display
-
This argument applies only with one- or two-dimensional data. The setting
"none"
will prevent any graphical output from being produced. In one dimensions, the default setting"line"
will produce the estimate. (For compatibility with earlier versions of the package, the setting"se"
will produce a variability band to show the variability, but not the bias, of the estimate. This should now be controlled by setting the separate parameterse
toTRUE
.) In two dimensions, the default setting"persp"
will produce a perspective plot of the estimate, while the settings"slice"
,"image"
and"rgl"
will produce slice (contour), image orrgl
plots. - eval.grid
-
logical flag which controls how the options
eval.points
are used for two-dimensional data. Ifeval.grid=TRUE
(default), evaluation is performed at points obtained by the cross-product of the two columns ofeval.points
. Ifeval.grid=FALSE
then evaluation is performed at points with coordinates specified by the coordinates ineval.points
. - eval.points
-
the points at which the density or the regression curve or surface estimate should be evaluated, for the values returned in the result of the function. This should be a vector for one-dimensional data and a two-column matrix for two-dimensional data. This argument does not apply with three-dimensional data.
- h.weights
-
a vector of weights which multiply the smoothing parameter used in the kernel function at each observation. This argument does not apply with three-dimensional data. Default value: 1.
- hmult
-
a factor which can be used to multiply the normal smoothing parameter before construction of the density estimate. Default value: 1.
- hull
-
a logical value which controls whether the estimate is evaluated and plotted only on grid points which fall within the convex hull of the data. When this argument is set to
FALSE
, evaluation and plotting take place at all grid points where the contribution from at least one kernel function is non-negligible. Both of these settings ensure that the estimate is not evaluated at points where there are no observations nearby. This argument applies only tosm.regression
andsm.discontinuity
in the case of two covariates. - lty
-
the line type used to plot the estimate. This argument applies only when the estimate is displayed as a curve or a contour. Default value: 1.
- method
-
the method used to select smoothing parameters. In density estimation the default is
"normal"
which uses a value which is asymptotically optimal for the normal distribution. Other possibilities are"cv"
for cross-validation and, for one-dimensional data only,"sj"
for the Sheather-Jones method.In nonparametric regression, the default is
"df"
which selects a smoothing parameters associated with the approximate degrees of freedom given in thedf
option. Other possibilities are"cv"
for cross-validation and"aicc"
for an AIC-based method proposed by Hurvich, Simonoff and Tsai. - nbins
-
the number of bins used in one-dimensional binning operations; in two-dimensional cases,
nbins
refers to the number of bins formed along each axis. Bins with 0 observed frequencies are ignored. Ifnbins=0
, binning is not performed; ifnbins=NA
(default), binning is switched on when the number of observations exceeds a certain threshold, which depends on the function. - nboot
-
number of samples generated in bootstraps. Default value: 100.
- ngrid
-
the number of points in the regular grid used to plot the estimate. For two- and three-dimensional data,
ngrid
refers to the number of points along the axis in each dimension. The same parameter is also used by a few other functions which perform some form of search (e.g.hcv
). Default value forsm.regression
: 50 and 20 for 1-, 2-dimensional data, respectively. Default value forsm.density
: 100, 50 and 20 for 1-, 2- and 3-dimensional data, respectively. - panel
-
a logical value which, when set to true, creates a panel which allows interactive control of
sm.regression
orsm.density
plots for one- or two-dimensional data. The panel can be used to alter the value of the smoothing parameter and control a variety of other settings. - panel.plot
-
a logical value which, when set to true (the default), places the plot inside the control panel (see the
panel
argument above), This creates a neater screen arrangement. - pch
-
the standard plotting character identified for data plotting. Default value: 1.
- period
-
a vector of length one or two identifying the period for covariates which are on a periodic scale. Periodic smoothing is implemented by local mean estimation, using a von Mises kernel function. Non-periodic covariates are identified by NA. Default value: NA.
- phi
-
the vertical rotation (in degrees) of perspective plots of estimate in the form of surfaces. Default value: 40.
- poly.index
-
an integer defining local constant (0) or local linear (1) smoothing. Default value: 1.
- positive
-
a logical value which indicates whether the data should be assumed to take positive values only, in
sm.density
. When this argument is set toTRUE
, a log transformation is applied to the data before construction of a density estimate. The result is transformed back to the original scale. This argument does not apply with three-dimensional data. Default value:FALSE
. - props
-
a vector defining the proportions of the data to be included within each contour in a slice plot, from two-dimensional data, or a contour surface plot, from three-dimensional data. In the three-dimensional case only the first element of the vector will be used. This argument does not apply to one-dimensional data. Default value:
c(75,50,25)
. - rugplot
-
logical flag which regulates whether a rugplot is superimposed to the density estimate, in the univariate case. Default value:
TRUE
. - se
-
logical flag which regulates whether a standard error information is added to the plot produced by
sm.regression
. If amodel
is specified, then these standard errors refer to the difference between this fitted model and the nonparametric regression estimate. Default value:TRUE
. - se.breaks
-
a numerical vector which defines the cut-points, on a standard error scale, for the assignment of colours when painting a regression surface with standard error information. Default value:
c(-3, -2, 3, 3)
. - show.script
-
logical flag which affects the behaviour of
sm.script
when this is called with non-empty argument. Ifshow.script=TRUE
(default) a window is opened to display the source code of the script. - size
-
an integer which defines the size of plotted points in
rgl
displays. The default value is2
. - structure.2d
-
the structure of the smoothing parameter in two-dimensional settings. The default is
"scaled"
, which uses the structure (h*sd(x[,1]), h*sd(x[,2])). Other possibilities are"separate"
, which uses (h1, h2), and"common"
which uses (h, h). The"common"
option may be particularly appropriate when the data have a spatial origin, where distances in each variable have the same meaning. Note that the"separate"
option is not available when"method"
is set to"df"
. - test
-
a logical flag controlling the production of a formal test, using the reference model as the null hypothesis. Default value:
TRUE
. - theta
-
the horizontal rotation (in degrees) of perspective plots of estimates in the form of surfaces. Default value: -30.
- verbose
-
regulates the amount of messages and other output printed out. If
verbose=0
only errors produce messages; ifverbose=1
(default value) warnings and the more relevant numerical output are printed ; ifverbose=2
more messages and more numerical output are printed. - xlab
-
the label attached to the x-axis.
- xlim
-
the range of the horizontal axis of the plot. This argument does not apply with three-dimensional data.
- yht
-
the upper limit of the vertical axis in a plot of a one-dimensional density estimate. The lower limit is always set to 0. This argument does not apply with two- or three-dimensional data.
- ylab
-
the label attached to the y-axis.
- ylim
-
the range of the vertical axis of the plot. This argument does not apply with three-dimensional data.
- zlab
-
the label attached to the z-axis (three-dimensional plots only).
- zlim
-
the range of the vertical axis when estimates are displayed as perspective plots.
Details
Arguments which are set by a function call will remain in effect until the end of the current S-plus session, unless overwritten by a subsequent call. In addition, they can be added as optional parameters of calls to specific functions of the sm
package; in this case, their effect is
limited to that function call.
See the documentation of specific functions for the list of options which are recognised by that function. Notice that some options are relevant only to some functions.
Value
a list with the updated values of the parameters. If the argument list is not empty, the returned list is invisible.
Examples
## Not run: sm.options(poly.index = 0)
# subsequent regression estimations will be performed using local means
# instead of local regression
#
sm.options(describe = FALSE)
# turns off typing documentation files of data loaded by `sm.script'
# (works from command-line)
#
## End(Not run)
Smooth principal components analysis
Description
This function calculates principal components in a manner which changes smoothly with a covariate. The smooth eigenvalues and eigenvector loadings can be plotted. A permutation test of equality of the components, both eigenvalues and eigenvectors, can be carried out.
Usage
sm.pca(x, Y, h, cor = TRUE, nperm = 100, pc = 1, ...)
Arguments
x |
either a vector of covariate values or a list object which is the output of a previous call to |
Y |
a matrix of responses whose rows correspond to the covariate values. |
h |
the smoothing parameter which controls the smoothness of estimation with respect to the covariate |
cor |
a logical value indicating whether the correlation, rather than covariance, matrix should be used. |
nperm |
the number of permutations used in the permutation test and graphical reference band. |
pc |
an integer value indicating the component to be plotted against the covariate. |
... |
other optional parameters are passed to the |
Details
Several further arguments may be set and these are passed to sm.options
. Relevant arguments for this function are display
("eigenvalues"
, "eigenvectors"
), ngrid
and df
. See link{sm.options}
for further details.
The smoothing is performed by the local constant kernel method and the smoothing parameter corresponds to the standard deviation of a normal kernel function. If h
is left unspecified then it is selected to correspond to the degrees of freedom set by the parameter df
.
The reference band for a constant eigenvalue is constructed from the upper and lower pointwise 2.5 percentiles of the smooth eigenvalue curves from the data with permuted covariate values. The p-value compares the observed value of the difference between the smoothed and constant eigenvalues, summed over the covariate grid in eval.points
, with the values generated from the permuted data.
In the eigenvector case, a reference band is computed from the percentiles of the curves from the permuted data, for each of the loadings. In order to plot all the loadings curves simultaneously, the locations where each curve lies inside its respective reference band are indicated by pale colour. The p-value compares the observed value of 1 - sum(e*e0)^2
, where e
and e0
are the eigenvectors under the smooth and constant scenarios (summed over the covariate grid), with the values generated from the permuted data. This test statistic differs from the one described in the Miller and Bowman (2012) reference below. It has been used as it conveniently handles the arbitrary sign of principal components.
When some components explain similar proportions of variance, the eigenvalues and eigenvectors can easily interchange, causing apparent sharp changes in the eigenvalue and eigenvector curves. It is difficult to track the components to avoid this.
Value
a list with the following components:
- xgrid
a vector of values on the covariate scale at which the smooth estimates are constructed.
- evals
a matrix whose columns give the smooth eigenvalues for each component at the covariate values.
- evecs
a three-dimensional array whose third dimension corresponds to the covariate values and whose second dimension indexes the smooth components.
- mhat
a matrix whose columns give the estimated smooth means for each dimension of
Y
at the covariate values.- var.explained
a matrix whose rows give the proportions of variance explained by each component at each covariate value.
- xlab
the label attached to the x-axis.
- h
the smoothing parameter used.
- x
the covariate values, after removal of missing cases.
- Y
the matrix of response values, after removal of missing cases.
- cor
a logical indicator of whether the correlation, rather than covariance, matrix is used in the construction of the eigenvalues and eigenvectors.
When a test or reference band is computed, the list has the additional components:
- nperm
the number of permutations used.
- evals.perm
the eigenvalues computed from the permuted data.
- evecs.perm
the eigenvectors computed from the permuted data.
When display contains "eigenvalues"
or "eigenvectors"
, the list has the additional components:
- p.values
the p-value for a test of constant eigenvalue for the component identified by
pc
.- p.vectors
the p-value for a test of constant eigenvectors for the component identified by
pc
.
When display contains "eigenvalues"
, the list has the additional component:
- band
a matrix whose two columns contain the boundaries of a reference band which indicates where the smooth eigenvalue curve should like if the hypothesis of no change in the eigenvalues with the covariate is correct.
When display contains "eigenvectors"
, the list has the additional components:
- xgrid.plot
a vector of values used for plotting the smooth eigenvectors.
- evecs.plot
a matrix whose rows contain the smooth eigenvectors at each value of
xgrid.plot
.- evecs.plot
a matrix whose columns contain the colours for the line segments in each smooth eigenvector component.
Side Effects
a plot on the current graphical device is produced, unless display="none"
References
Miller, C. and Bowman, A.W. (2012). Smooth principal components for investigating changes in covariances over time. Applied Statistics 61, 693–714.
See Also
Examples
## Not run:
Y <- log(as.matrix(aircraft[ , -(1:2)]))
year <- aircraft$Yr
h <- h.select(year, Y[ , 1], method = "df", df = 4)
spca <- sm.pca(year, Y, h, display = "none")
sm.pca(year, Y, h, display = "eigenvalues")
sm.pca(year, Y, h, display = "eigenvectors", ylim = c(-1, 1))
# The following code shows how the plots can be redrawn from the returned object
spca <- sm.pca(year, Y, h, display = "eigenvalues")
spca <- sm.pca(year, Y, h, display = "eigenvectors", ylim = c(-1, 1))
with(spca, {
ylim <- range(evals[ , 1], band)
plot(xgrid, evals[ , 1], type = "n", ylab = "Variance", ylim = ylim)
polygon(c(xgrid, rev(xgrid)), c(band[ , 1], rev(band[ , 2])),
col = "lightgreen", border = NA)
lines(xgrid, evals[ , 1], col = "red")
})
with(spca, {
pc <- 1
plot(range(xgrid.plot), range(evecs.plot), type = "n",
xlab = "x", ylab = "PC loadings")
for (i in 1:ncol(Y))
segments(xgrid.plot[-length(xgrid.plot)],
evecs.plot[-nrow(evecs.plot), i],
xgrid.plot[-1], evecs.plot[-1, i],
col = col.plot[ , i], lty = i)
})
## End(Not run)
Nonparametric Poisson regression
Description
This function estimates the regression curve using the local likelihood approach for a vector of Poisson observations and an associated vector of covariate values.
Usage
sm.poisson(x, y, h, ...)
Arguments
x |
vector of the covariate values |
y |
vector of the response values; they must be nonnegative integers. |
h |
the smoothing parameter; it must be positive. |
... |
other optional parameters are passed to the |
Details
see Sections 3.4 and 5.4 of the reference below.
Value
A list containing vectors with the evaluation points, the corresponding probability estimates, the linear predictors, the upper and lower points of the variability bands and the standard errors on the linear predictor scale.
Side Effects
graphical output will be produced, depending on the value of the
display
option.
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
See Also
sm.binomial
, sm.binomial.bootstrap
,
binning
, glm
Examples
with(muscle, {
TypeI <- TypeI.R+ TypeI.P+TypeI.B
sm.poisson(x=log(TypeI), y=TypeII, h=0.25,display="se")
sm.poisson(x=log(TypeI), y=TypeII, h=0.75, col=2, add=TRUE)
})
Bootstrap goodness-of-fit test for a Poisson regression model
Description
This function is associated with sm.poisson
for the underlying
fitting procedure. It performs a Pseudo-Likelihood Ratio Test for the
goodness-of-fit of a standard parametric Poisson regression of specified
degree
in the covariate x
.
Usage
sm.poisson.bootstrap(x, y, h, degree = 1,
fixed.disp = FALSE, intercept = TRUE, ...)
Arguments
x |
vector of the covariate values |
y |
vector of the response values; they must be nonnegative integers. |
h |
the smoothing parameter; it must be positive. |
degree |
specifies the degree of the fitted polynomial in |
fixed.disp |
if |
intercept |
|
... |
additional parameters passed to |
Details
see Section 5.4 of the reference below.
Value
a list containing the observed value of the Pseudo-Likelihood Ratio Test statistic, its observed p-value as estimated via the bootstrap method, and the vector of estimated dispersion parameters when this value is not forced to be 1.
Side Effects
Graphical output representing the bootstrap samples is produced on the current graphical device. The estimated dispersion parameter, the value of the test statistic and the observed significance level are printed.
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
See Also
sm.poisson
, sm.binomial.bootstrap
Examples
## takes a while: extend sm.script(muscle)
with(muscle, {
TypeI <- TypeI.P + TypeI.R + TypeI.B
sm.poisson.bootstrap(log(TypeI), TypeII, h = 0.5)
})
Nonparametric regression with one or two covariates.
Description
This function creates a nonparametric regression estimate from data
consisting of a single response variable and one or two covariates.
In two dimensions a perspective, image (image
), contour (slice
)
or rgl
plot of the estimated regression surface is produced.
A number of other features of the construction of the estimate, and of
its display, can be controlled.
If the rpanel
package is available, an interactive panel can be activated
to control various features of the plot.
Usage
sm.regression(x, y, h, design.mat = NA, model = "none", weights = NA,
group = NA, ...)
Arguments
x |
a vector, or two-column matrix, of covariate values. |
y |
a vector of responses. |
h |
a vector of length 1 or 2 giving the smoothing parameter. A normal kernel
function is used and |
design.mat |
the design matrix used to produce |
model |
a character variable which defines a reference model. The settings
|
weights |
a vector which allows the kernel functions associated with the observations
to take different weights. This is useful, in particular, when different
observations have different precisions.
The normal usage of this parameter is to associate observations with
frequencies; if the |
group |
a vector of groups indicators (numeric or character values) or
a factor. If this argument is used then the data are passed to the |
... |
other optional parameters are passed to the |
Details
When display
is set to "persp"
or "rgl"
, a number of
graphical options are available. By setting the col
parameter to
"height"
or "se"
, the surface will be painted by colours to
reinforce the perception of height or indicate the relative sizes of the
standard errors respectively. When model
is not "none"
,
the colour coding refers to the number of standard errors which separate
the smooth regression surface and the nominated model at each position.
The parameter "se.breaks"
, whose default value is c(-3, -2, 3, 3)
can then be used to set the colour ranges. In this case, col.palette
must be set to a list of colours whose length is one greater than the length
of the cut-points in "se.breaks"
. If this is not the case, the
default colour palette
rev(rainbow(length(opt$se.breaks) + 1, start = 0/6, end = 4/6))
.
If the argument col
is not set then surface painting will be determined
by the setting of se
. If neither is set then colour painting will be
activated by default if model != "none"
. (In this latter case, the
argument band
, retained from earlier versions for compatibility, will
also be examined.)
When display
is set to "rgl"
, some additional parameters
can be used to control details of the plot. Transparency can be set by
alpha
, which lies between 0
and 1
. When alpha
is set to a vector of length two, the first component refers to the surface
panels and the second to the surface mesh. Setting a component of alpha
to 0
will remove the corresponding feature from the plot. col.mesh
,
whose valid values match those of col
, controls the colour of the surface
mesh. The logical parameter lit
has the same meaning as in the rgl
package; see material3d
.
When panel
is set to "TRUE"
, an interactive control panel is
created if the rpanel
package is available.
If a covariate is on a cyclical scale, this can be incorporated by setting
the period
argument to a vector (of length 1 or 2) whose components give
the values of the periods, or NA if the covariate is not periodic.
See Chapters 3, 4 and 5 of the first reference below for the details of the construction of the estimate and its standard error. The second reference gives further details and examples of surface painting.
Value
a list containing the values of the estimate at the evaluation points,
the smoothing parameter and the smoothing parameter weights.
If a reference model has been specified and test
set to
TRUE
, then the p-value of the test
is also returned. When there is only one covariate, the weights associated
with different observations, an estimate of the error standard deviation and
the standard error of the estimate are also returned. If a reference model
has been specified, this standard error refers to the comparison between
the estimate and the reference model, and the values defining the reference
model are also returned.
If an rgl
display is used, then the indices of the surface and lines
used to create the display are returned.
Side Effects
a plot on the current graphical device is produced, unless the option
display="none"
is set.
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
Bowman, A.W. (2006). Comparing nonparametric surfaces. Statistical Modelling, 6, 279-299.
See Also
hcv
, sm
, sm.ancova
,
sm.binomial
, sm.poisson
,
sm.regression.autocor
, sm.survival
,
sm.options
, sm.surface3d
Examples
with(trawl, {
Zone92 <- (Year == 0 & Zone == 1)
Position <- cbind(Longitude - 143, Latitude)
dimnames(Position)[[2]][1] <- "Longitude - 143"
par(mfrow = c(2, 2))
sm.regression(Longitude, Score1, method = "aicc", col = "red",
model = "linear")
sm.regression(Position[Zone92, ], Score1[Zone92], display = "image",
theta = 120)
sm.regression(Position[Zone92, ], Score1[Zone92], df = 12, col = "se",
theta = 120)
sm.regression(Position[Zone92, ], Score1[Zone92], df = 12, col = "se",
model = "linear", theta = 120)
par(mfrow = c(1, 1))
})
# sm.regression(Position[Zone92, 2:1], Score1[Zone92], display = "rgl", df = 12)
# sm.regression(Position[Zone92, 2:1], Score1[Zone92], display = "rgl", df = 12,
# alpha = c(0.9, 1), col = "se", model = "linear")
# sm.regression(Position[Zone92, 1], Score1[Zone92], panel = TRUE)
# sm.regression(Position[Zone92, ], Score1[Zone92], panel = TRUE)
# sm.regression(Position[Zone92, ], Score1[Zone92], panel = TRUE, display = "rgl")
Nonparametric regression with autocorrelated errors
Description
This function estimates nonparametrically the regression function
of y
on x
when the error terms are serially correlated.
Usage
sm.regression.autocor(x = 1:n, y, h.first, minh, maxh, method = "direct", ...)
Arguments
y |
vector of the response values |
h.first |
the smoothing parameter used for the initial smoothing stage. |
x |
vector of the covariate values; if unset, it is assumed to
be |
minh |
the minimum value of the interval where the optimal smoothing parameter is searched for (default is 0.5). |
maxh |
the maximum value of the interval where the optimal smoothing parameter is searched for (default is 10). |
method |
character value which specifies the optimality criterium adopted;
possible values are |
... |
other optional parameters are passed to the |
Details
see Section 7.5 of the reference below.
Value
a list as returned from sm.regression called with the new value of
smoothing parameter, with an additional term $aux
added which contains
the initial value h.first
, the estimated curve using h.first
,
the autocorrelation function of the residuals from the initial fit,
and the residuals.
Side Effects
a new suggested value for h
is printed; also, if the parameter display
is not equal to "none"
, graphical output is produced on the current
graphical device.
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
See Also
sm.regression
, sm.autoregression
Nonparametric analysis of repeated measurements data
Description
This function estimates nonparametrically the mean profile from a matrix
y
which is assumed to contain repeated measurements (i.e. longitudinal
data) from a set of individuals.
Usage
sm.rm(Time, y, minh = 0.1, maxh = 2, optimize = FALSE,
rice.display = FALSE, ...)
Arguments
y |
matrix containing the values of the response variable, with rows associated to individuals and columns associated to observation times. |
Time |
a vector containing the observation times of the response variable, assumed
to be the same for all individuals of matrix |
minh |
the minimum value of the interval where the optimal value of the smoothing parameter is searched according to the modified Rice criterion. See reference below for details. |
maxh |
the maximum value of the above interval. |
optimize |
Logical value, default is |
rice.display |
If this set to |
... |
other optional parameters are passed to the
|
Details
see Section 7.4 of the reference below.
Value
a list containing the returned value produced by sm.regression
when
smoothing the mean response value at each given observation time,
with an extra component $aux
added to the list.
This additional component is a list itself containing the mean value at each
observation time, the residual variance of the residuals from the estimated
regression curve, the autocorrelation function of the residuals, and
the value h
of the chosen smoothing parameter.
Side Effects
if the parameter display is not set to "none"
, a plot of the estimated
regression curve is produced;
other aspects are controlled by the optional parameters (...
).
If rice.display=TRUE
, a plot of the modified Rice criterion is shown.
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
See Also
sm.regression
, sm.regression.autocor
, optim
Examples
sm.rm(y=as.matrix(citrate), display.rice=TRUE)
#
with(dogs, {
Time <- seq(1,13,by=2)
gr1 <- as.matrix(dogs[dogs$Group==1,2:8])
plot(c(1,13), c(3,6),xlab="time", ylab="potassium", type="n")
sm1 <- sm.rm(Time, gr1, display="se", add=TRUE)
})
Running a script associated to the sm library
Description
This is a utility function to run scripts, usually those associated with the book described below.
Usage
sm.script(name, path)
Arguments
name |
the name of the file containing the code; a |
path |
the name of the path where to look for the script. If |
Details
This utility allows the illustrations of the reference below to be reproduced
easily, since each of them has an associated script. The display of the
script file itself is controlled by the setting of the logical variable
show.script
in sm.options
. This is set to TRUE
by default.
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
See Also
sm
,
Examples
sm.script()
sm.script(speed)
Estimation of the error standard deviation in nonparametric regression.
Description
This function estimates the error standard deviation in nonparametric regression with one or two covariates.
Usage
sm.sigma(x, y, rawdata = NA, weights = rep(1, length(y)),
diff.ord = 2, ci = FALSE, model = "none", h = NA, ...)
Arguments
x |
a vector or two-column matrix of covariate values. |
y |
a vector of responses. |
rawdata |
a list containing the output from a binning operation.
This argument is used by |
weights |
a list of frequencies associated with binned data.
This argument is used by |
diff.ord |
an integer value which determines first (1) or second (2) differencing in the estimation of sigma. |
ci |
a logical value which controls whether a confidence interval is produced. |
model |
a character variable. If this is set to |
h |
a vector of length two defining a smoothing parameter to be used in the test of constant variance. |
... |
other optional parameters are passed to the |
Details
see the reference below.
Value
a list containing the estimate and, in the two covariate case, a
matrix which can be used by the function sm.sigma2.compare
,
pseudo-residuals and, if appropriate, a confidence interval and
a p-value for the test of constant variance.
Side Effects
none.
References
Bock, M., Bowman, A.W. & Ismail, B. (2007). Estimation and inference for error variance in bivariate nonparametric regression. Statistics & Computing, to appear.
See Also
Examples
## Not run:
with(airquality, {
x <- cbind(Wind, Temp)
y <- Ozone^(1/3)
group <- (Solar.R < 200)
sig1 <- sm.sigma(x[ group, ], y[ group], ci = TRUE)
sig2 <- sm.sigma(x[!group, ], y[!group], ci = TRUE)
print(c(sig1$estimate, sig1$ci))
print(c(sig2$estimate, sig2$ci))
print(sm.sigma(x[ group, ], y[ group], model = "constant", h = c(3, 5))$p)
print(sm.sigma(x[!group, ], y[!group], model = "constant", h = c(3, 5))$p)
print(sm.sigma2.compare(x[group, ], y[group], x[!group, ], y[!group]))
})
## End(Not run)
Comparison across two groups of the error standard deviation in nonparametric regression with two covariates.
Description
This function compares across two groups, in a hypothesis test, the error standard deviation in nonparametric regression with two covariates.
Usage
sm.sigma2.compare(x1, y1, x2, y2)
Arguments
x1 |
a two-column matrix of covariate values for group 1. |
y1 |
a vector of responses for group 1. |
x2 |
a two-column matrix of covariate values for group 2. |
y2 |
a vector of responses for group 2. |
Details
see the reference below.
Value
a p-value for the test of equality of standard deviations.
Side Effects
none.
References
Bock, M., Bowman, A.W. & Ismail, B. (2007). Estimation and inference for error variance in bivariate nonparametric regression. Statistics & Computing, to appear.
See Also
Examples
## Not run:
with(airquality, {
x <- cbind(Wind, Temp)
y <- Ozone^(1/3)
group <- (Solar.R < 200)
sig1 <- sm.sigma(x[ group, ], y[ group], ci = TRUE)
sig2 <- sm.sigma(x[!group, ], y[!group], ci = TRUE)
print(c(sig1$estimate, sig1$ci))
print(c(sig2$estimate, sig2$ci))
print(sm.sigma(x[ group, ], y[ group], model = "constant", h = c(3, 5))$p)
print(sm.sigma(x[!group, ], y[!group], model = "constant", h = c(3, 5))$p)
print(sm.sigma2.compare(x[group, ], y[group], x[!group, ], y[!group]))
})
## End(Not run)
Nonparametric density estimation for spherical data.
Description
This function creates a density estimate from data which can be viewed as lying on the surface of a sphere. Directional data form a principal example. The data are displayed in spherical form and a density estimate may be superimposed. The angle of view may be altered. An interactive panel is available to control some features of the estimate and the display. Only modest amounts of data may be used. The limit will depend on the memory available.
Usage
sm.sphere(lat, long, kappa = 20, hidden = FALSE, sphim = FALSE,
addpoints = FALSE, ...)
Arguments
lat |
a vector giving the latitude component of the data in degrees from the equator. |
long |
a vector giving the longitude component of the data in degrees east. |
kappa |
the smoothing parameter used to construct the density estimate. The kernel
function is a Fisher distribution and |
a logical value which, when set to | |
sphim |
a logical value which controls whether a density estimate is constructed and displayed on the sphere in image form. |
addpoints |
a logical value which controls whether the data points are added to the plot of the density estimate. |
... |
arguments for |
Details
see Section 1.5 of the reference below.
Value
a list containing the value of the smoothing parameter and the rotation angles of the displayed plot.
Side Effects
none.
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
See Also
Examples
lat <- rnorm(50, 10, 15)
long <- c(rnorm(25, 300, 15), rnorm(25, 240, 15))
par(mfrow=c(1,2))
sm.sphere(lat, long)
sm.sphere(lat, long, sphim=TRUE, kappa=15)
par(mfrow=c(1,1))
Adding a regression surface to an rgl plot.
Description
This function adds a regression surface, defined by a matrix of heights
at a regular grid of values of two covariates, to an rgl
plot.
Missing values can be accommodated.
Usage
sm.surface3d(eval.points, surf, scaling,
col = "green", col.mesh = "black",
alpha = 0.7, alpha.mesh = 1, lit = TRUE, ...)
Arguments
eval.points |
if this is a two-column matrix then each column defines the marginal grids of covariate values. Alternatively, a list with two components can also be used to handle cases where the grids are of different size. |
surf |
a matrix of heights corresponding to the grid of covariate values. NAs are allowed. |
scaling |
a function to define the scaling for the |
col |
the colour of the surface. If |
col.mesh |
the colour of the surface mesh. If |
alpha |
the transparency of the filled triangles defining the surface. Setting
this to |
alpha.mesh |
the transparency of the lines drawn across the regular grid of covariate
values. Setting this to |
lit |
a logical variable which controls whether the |
... |
other optional parameters which are passed to |
Details
the principal motivation for this function is that is can handle missing
data in regression surfaces. In particular, it can be used to plot the
results of applying sm.regression
. In addition, the function can
be used to build up more complex plots by adding successive surfaces.
Value
a vector of length 2 containing the ids of the filled surface and lines
added to the rgl
plot.
Side Effects
a surface is added to the rgl
plot.
See Also
Examples
with(trawl, {
Zone93 <- (Year == 1 & Zone == 1)
Position <- cbind(Longitude - 143, Latitude)
model1 <- sm.regression(Position[Zone93,], Score1[Zone93],
h= c(0.1, 0.1), display = "rgl", xlab="Longitude - 143")
model2 <- sm.regression(Position[Zone93,], Score1[Zone93],
h= c(0.2, 0.2), display = "none")
sm.surface3d(model2$eval.points, model2$est, model1$scaling, col = "red")
})
Nonparametric regression with survival data.
Description
This function creates a smooth, nonparametric estimate of the quantile of the distribution of survival data as a function of a single covariate. A weighted product-limit estimate of the survivor function is obtained by smoothing across the covariate scale. A small amount of smoothing is then also applied across the survival time scale in order to achieve a smooth estimate of the quantile.
Usage
sm.survival(x, y, status, h , hv = 0.05, p = 0.5, status.code = 1, ...)
Arguments
x |
a vector of covariate values. |
y |
a vector of survival times. |
status |
an indicator of a complete survival time or a censored value. The value of
|
h |
the smoothing parameter applied to the covariate scale. A normal kernel
function is used and |
hv |
a smoothing parameter applied to the weighted to the product-limit estimate derived from the smoothing procedure in the covariate scale. This ensures that a smooth estimate is obtained. |
p |
the quantile to be estimated at each covariate value. |
status.code |
the value of |
... |
other optional parameters are passed to the |
Details
see Section 3.5 of the reference below.
Value
a list containing the values of the estimate at the evaluation points and the values of the smoothing parameters for the covariate and survival time scales.
Side Effects
a plot on the current graphical device is produced, unless the option
display="none"
is set.
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
See Also
Examples
x <- runif(50, 0, 10)
y <- rexp(50, 2)
z <- rexp(50, 1)
status <- rep(1, 50)
status[z<y] <- 0
y <- pmin(z, y)
sm.survival(x, y, status, h=2)
Nonparametric density estimation of stationary time series data
Description
This function estimates the density function of a time series x
,
assumed to be stationary. The univariate marginal density is estimated
in all cases; bivariate densities of pairs of lagged values are estimated
depending on the parameter lags
.
Usage
sm.ts.pdf(x, h = hnorm(x), lags, maxlag = 1, ask = TRUE)
Arguments
x |
a vector containing a time series |
h |
bandwidth |
lags |
for each value, |
maxlag |
if |
ask |
if |
Details
see Section 7.2 of the reference below.
Value
a list of two elements, containing the outcome of the estimation of
the marginal density and the last bivariate density, as produced by
sm.density
.
Side Effects
plots are produced on the current graphical device.
References
Bowman, A.W. and Azzalini, A. (1997). Applied Smoothing Techniques for Data Analysis: the Kernel Approach with S-Plus Illustrations. Oxford University Press, Oxford.
See Also
Examples
with(geyser, {
sm.ts.pdf(geyser$duration, lags=1:2)
})
Confidence intervals and tests based on smoothing an empirical variogram.
Description
This function constructs an empirical variogram, using the robust form of construction based on square-root absolute value differences of the data. Flexible regression is used to assess a variety of questions about the structure of the data used to construct the variogram, including independence, isotropy and stationarity. Confidence bands for the underlying variogram, and reference bands for the independence, isotropy and stationarity models, can also be constructed under the assumption that the errors in the data are approximately normally distributed.
Usage
sm.variogram(x, y, h, df.se = "automatic", max.dist = NA, n.zero.dist = 1,
original.scale = TRUE, varmat = FALSE, ...)
Arguments
x |
a vector or two-column matrix of spatial location values. |
y |
a vector of responses observed at the spatial locations. |
h |
a smoothing parameter to be used on the distance scale. A normal kernel
function is used and |
df.se |
the degrees of freedom used when smoothing the empirical variogram to estimate standard errors. The default value of "automatic" selects the degrees of smoothing described in the Bowman and Crujeiras (2013) reference below. |
max.dist |
this can be used to constrain the distances used in constructing the variogram. The default is to use all distances. |
n.zero.dist |
an integer value which sets the lower limit on the number of pairs of data points with distance zero (in other words the number of pairs whose positions are identical) to trigger a separate variogram bin for zero distance. Repeat observations at the same location can be informative about the variance of the underlying process. The default for This parameter has no effect when |
original.scale |
a logical value which determines whether the plots are constructed on the original variogram scale (the default) or on the square-root absolute value scale on which the calculations are performed. |
varmat |
a logical value which determines whether the variance matrix of the estimated variogram is returned. |
... |
other optional parameters are passed to the An important parameter here is Other relevant parameters are
|
Details
The reference below describes the statistical methods used in the function.
Note that, apart from the simple case of the independence model, the calculations required are extensive and so the function can be slow.
The display
argument has a special meaning for
this function. Its default value is "binned"
, which plots the
binned version of the empirical variogram. As usual, the value "none"
will suppress the graphical display. Any other value will lead to a plot
of the individual differences between all observations. This will lead
to a very large number of plotted points, unless the dataset is small.
Value
A list with the following components:
sqrtdiff , distance |
the raw differences and distances |
sqrtdiff.mean , distance.mean |
the binned differences and distances |
weights |
the frequencies of the bins |
estimate |
the values of the estimate at the evaluation points |
eval.points |
the evaluation points |
h |
the value of the smoothing parameter used |
ibin |
an indicator of the bin in which the distance between each pair of observations was placed |
ipair |
the indices of the original observations used to construct each pair |
When model = "isotropic"
or model = "stationary"
the following components may also be returned, depending on the arguments passed in ... or the settings in sm.options
:
p |
the p-value of the test |
se |
the standard errors of the binned values (if the argument |
se.band |
when an independence model is examined, this gives the standard error of the difference between the smooth estimate and the mean of all the data points, if a reference band has been requested |
V |
the variance matrix of the binned variogram. When |
sdiff |
the standardised difference between the estimate of the variogram and the reference model, evaluated at |
levels |
the levels of standardised difference at which contours are drawn in the case of |
Side Effects
a plot on the current graphical device is produced, unless the option display = "none"
is set.
References
Diblasi, A. and Bowman, A.W. (2001). On the use of the variogram for checking independence in a Gaussian spatial process. Biometrics, 57, 211-218.
Bowman, A.W. and Crujeiras, R.M. (2013). Inference for variograms. Computational Statistics and Data Analysis, 66, 19-31.
See Also
Examples
## Not run:
with(coalash, {
Position <- cbind(East, North)
sm.options(list(df = 6, se = TRUE))
par(mfrow=c(2,2))
sm.variogram(Position, Percent, original.scale = FALSE, se = FALSE)
sm.variogram(Position, Percent, original.scale = FALSE)
sm.variogram(Position, Percent, original.scale = FALSE, model = "independent")
sm.variogram(East, Percent, original.scale = FALSE, model = "independent")
par(mfrow=c(1,1))
})
# Comparison of Co in March and September
with(mosses, {
nbins <- 12
vgm.m <- sm.variogram(loc.m, Co.m, nbins = nbins, original.scale = TRUE,
ylim = c(0, 1.5))
vgm.s <- sm.variogram(loc.s, Co.s, nbins = nbins, original.scale = TRUE,
add = TRUE, col.points = "blue")
trns <- function(x) (x / 0.977741)^4
del <- 1000
plot(vgm.m$distance.mean, trns(vgm.m$sqrtdiff.mean), type = "b",
ylim = c(0, 1.5), xlab = "Distance", ylab = "Semi-variogram")
points(vgm.s$distance.mean - del, trns(vgm.s$sqrtdiff.mean), type = "b",
col = "blue", pch = 2, lty = 2)
plot(vgm.m$distance.mean, trns(vgm.m$sqrtdiff.mean), type = "b",
ylim = c(0, 1.5), xlab = "Distance", ylab = "Semi-variogram")
points(vgm.s$distance.mean - del, trns(vgm.s$sqrtdiff.mean), type = "b",
col = "blue", pch = 2, lty = 2)
segments(vgm.m$distance.mean, trns(vgm.m$sqrtdiff.mean - 2 * vgm.m$se),
vgm.m$distance.mean, trns(vgm.m$sqrtdiff.mean + 2 * vgm.m$se))
segments(vgm.s$distance.mean - del, trns(vgm.s$sqrtdiff.mean - 2 * vgm.s$se),
vgm.s$distance.mean - del, trns(vgm.s$sqrtdiff.mean + 2 * vgm.s$se),
col = "blue", lty = 2)
mn <- (vgm.m$sqrtdiff.mean + vgm.s$sqrtdiff.mean) / 2
se <- sqrt(vgm.m$se^2 + vgm.s$se^2)
plot(vgm.m$distance.mean, trns(vgm.m$sqrtdiff.mean), type = "n",
ylim = c(0, 1.5), xlab = "Distance", ylab = "Semi-variogram")
polygon(c(vgm.m$distance.mean, rev(vgm.m$distance.mean)),
c(trns(mn - se), rev(trns(mn + se))),
border = NA, col = "lightblue")
points(vgm.m$distance.mean, trns(vgm.m$sqrtdiff.mean))
points(vgm.s$distance.mean, trns(vgm.s$sqrtdiff.mean), col = "blue", pch = 2)
vgm1 <- sm.variogram(loc.m, Co.m, nbins = nbins, varmat = TRUE,
display = "none")
vgm2 <- sm.variogram(loc.s, Co.s, nbins = nbins, varmat = TRUE,
display = "none")
nbin <- length(vgm1$distance.mean)
vdiff <- vgm1$sqrtdiff.mean - vgm2$sqrtdiff.mean
tstat <- c(vdiff %*% solve(vgm1$V + vgm2$V) %*% vdiff)
pval <- 1 - pchisq(tstat, nbin)
print(pval)
})
# Assessing isotropy for Hg in March
with(mosses, {
sm.variogram(loc.m, Hg.m, model = "isotropic")
})
# Assessing stationarity for Hg in September
with(mosses, {
vgm.sty <- sm.variogram(loc.s, Hg.s, model = "stationary")
i <- 1
image(vgm.sty$eval.points[[1]], vgm.sty$eval.points[[2]], vgm.sty$estimate[ , , i],
col = topo.colors(20))
contour(vgm.sty$eval.points[[1]], vgm.sty$eval.points[[2]], vgm.sty$sdiff[ , , i],
col = "red", add = TRUE)
})
## End(Not run)
Mackerel data from a Spanish survey
Description
These data were recorded by a Spanish survey, as part of a multi-country survey of the abundance of mackerel eggs off the coast of north-western Europe, in 1992.
The variables are:
Density | egg density |
smack.lat | latitude of sampling position |
smack.long | longitude of sampling position |
smack.depth | bottom depth |
Temperature | surface temperature |
D200 | distance from the 200m depth contour line |
Background to the survey and the data are provided by Watson et al. (1992), Priede and Watson (1993) and Priede et al (1995). Borchers et al (1997) describe an analysis of the data.
Survival times from the Stanford Heart Transplant Study
Description
These data refer to the survival times of patients of different ages from the Stanford Heart Transplant Study.
The variables are:
Log.time | egg density |
Status | latitude of sampling position |
Age | longitude of sampling position |
Source: Miller & Halpern (1980). Regression with censored data. Biometrika 69, 521-531.
Tephra layer
Description
These data record the percentages of aluminium oxide found in samples from a tephra layer resulting from a volcanic eruption in Iceland around 3500 years ago.
The variables are:
Al2O3 | percentage of aluminium oxide |
The data were collected by A.Dugmore. The geological background to the data is given by Dugmore et al (1992), Geochemical stability of finegrained silicic tephra in Iceland and Scotland, J.Quatern.Sci. 7, 173-83.
Trawl data from the Great Barrier Reef
Description
These data refer to a survey of the fauna on the sea bed lying between the coast of northern Queensland and the Great Barrier Reef. The sampling region covered a zone which was closed to commercial fishing, as well as neighbouring zones where fishing was permitted.
The variables are:
Zone | an indicator for the closed (1) and open (0) zones |
Year | an indicator of 1992 (0) or 1993 (1) |
Latitude | latitude of the sampling position |
Longitude | longitude of the sampling position |
Depth | bottom depth |
Score1 | catch score 1 |
Score2 | catch score 2 |
The details of the survey and an analysis of the data are provided by Poiner et al. (1997), The effects of prawn trawling in the far northern section of the Great Barrier Reef, CSIRO Division of Marine Research, Queensland Dept. of Primary Industries.
Potassium cyanate and trout eggs
Description
These data were collected in a toxocological experiment conducted at the University of Waterloo. Different concentrations of potassium cyanate were applied to vials of trout eggs. The eggs in half of the vials were allowed to water-harden before the toxicant was applied.
The variables are:
Concentr | toxicant concentration |
Trouts | number of trout eggs |
Dead | number of eggs which died |
Insert | an indicator of whether the eggs were allowed to water-harden |
Source: O'Hara Hines & Carter (1993). Improved added variable and partial residual plots for the detection of influential observations in generalized linear models. Applied Statistics 42, 3-20.
The data are also reported by Hand et al. (1994), A Handbook of Small Data Sets, data set no.418.
Yield-density relationship for White Imperial Spanish onion plants
Description
These data were collected in a study of the relationship between the yield of White Imperial Spanish onion plants and the density of planting.
The variables are:
Density | density of planting (plants/m^2) |
Yield | yield (g/plant) |
Locality | a code to indicate Purnong Landing (1) or Virginia (2) |
The data were collected by I.S.Rogers (South Australian Dept. of Agriculture & Fisheries). They are listed in Ratkowsky (1983), Nonlinear Regression Modeling. Dekker, New York.
Human parasitic worm infections
Description
These data record the occurrence of a human parasitic worm infection in residents of a rural community in China.
The variables are:
Age | age of the resident |
Infection | presence (1) or absence (0) of infection |
Sex | male (1) or female (2) |
The background to the data, and an analysis, are described by Weidong et al. (1996), Ascaris, people and pigs in a rural community of Jiangxi province, China, Parasitology 113, 545-57.