Article Text

Download PDFPDF

Prospective cohort study to assess rates of contagious disease in pre-weaned UK dairy heifers: management practices, passive transfer of immunity and associated calf health
  1. Kate F Johnson,
  2. Natalie Chancellor,
  3. Charlotte C Burn and
  4. D Claire Wathes
  1. Department of Pathobiology and Population Sciences, Royal Veterinary College, Hatfield, UK
  1. Correspondence to Professor D Claire Wathes; dcwathes{at}


Dairy calves are vulnerable to infectious diseases, particularly diarrhoea and bovine respiratory disease (BRD), causing mortality and reducing welfare and growth. A prospective cohort study was performed on 11 UK dairy farms to determine the underlying causes for calf disease. This first paper describes the incidence, timing and duration of infectious disease, mortality rates, passive transfer of immunity and key management practices that may contribute to disease incidence. Heifer calves were recruited in the first week of life (n=492) and a blood sample taken to measure IgG and total protein (TP). Each animal was examined weekly for nine weeks using a standardised health scoring system. Recruitment of calves occurred between August and February. Four farms provided supplementary colostrum to more than 75 per cent of calves born, whereas on the remainder only 0 to 19 per cent were supplemented. Mean serum IgG and TP were 19.0±10 and 56.7±10.3 mg/ml respectively, with 20.7 per cent (95CI: 17.2 to 24.7 per cent) of all calves classified as having failure of passive transfer (IgG <10 mg/ml). The overall preweaning mortality rate was 4.5 per cent. (95 per cent CI: 2.9 to 6.8 per cent). During this period,48.2 per cent of all calves (range 24.1 to 74.4 per cent between farms) were diagnosed with diarrhoea and 45.9 per cent (range 20.4 to 77.8 per cent) with BRD. The incidence rates were 7.8 cases of diarrhoea and 10.1 cases of BRD per 100 calf weeks at risk, respectively. Rates of infectious disease were therefore high despite relatively good passive transfer.

  • calves
  • dairy cattle
  • diarrhoea
  • health
  • respiratory disease
  • passive immunity

This is an Open Access article distributed in accordance with the terms of the Creative Commons Attribution (CC BY 4.0) license, which permits others to distribute, remix, adapt and build upon this work, for commercial use, provided the original work is properly cited. See:

View Full Text

Statistics from


Most mortality and contagious disease in dairy heifers occurs during the milk feeding period.1–3 Reported mortality rates in young calves (up to age two–six months depending on the study and excluding perinatal losses) were between 2.4 per cent and 11.7 per cent, with diarrhoea and bovine respiratory disease (BRD) listed as the major causes in this age group (reviewed by Johnson and others).4 Death rates of calves on individual farms within studies varied widely from 0 to 30 per cent. As some farms were able to achieve less than one per cent overall mortality, this suggested that deaths on other farms may be preventable with improved husbandry using existing strategies.5 6

Calf mortality is also indicative of underlying problems. Sick animals have poor welfare and are unable to express normal behaviours.7 There are associated veterinary expenses and an economic cost of reduced growth rates.8 Juvenile disease may compromise development in surviving heifers and prevent them from reaching rearing targets.9–11 For example, first parity was delayed by a median of six months in heifers that had BRD in the first three months of life12 and calves that had four episodes of BRD before first calving had 1.87±0.14 greater odds of failing to complete their first lactation compared with healthy calves.13 Such animals are also unlikely to recoup their rearing costs.

BRD and diarrhoea can be diagnosed and recorded by several different systems, and several studies have shown that this can influence the disease rates reported. EU farmers must keep a medicines book, although this does not include rehydration treatments for diarrhoea. Comparisons between farmer diagnosis with weekly clinical examinations by veterinarians showed only moderate agreement, with Sivula and others14 reporting a kappa statistic of 0.47 with respect to BRD. This is a test of interobserver agreement, measured on a scale of 0, chance to 1, perfect agreement. Virtala and others10 found that pneumonia diagnosed by a veterinarian was significantly associated with a reduced bodyweight gain in calves (P<0.01), whereas this relationship was not significant when farmer diagnosis was used instead (P=0.47). They attributed this difference to false-positive diagnoses by the farmer. On the other hand, only 47 per cent of cases of respiratory disease detected by farmers were treated with antibiotics, highlighting frequently missed treatment opportunities.15 These findings call into question the reliability of disease diagnostic information supplied by producers. The format in which data are reported can also vary. Cumulative incidence, defined as the total number of new cases during the study period, is widely used. This does not, however, account for duration of disease and repeated cases in an individual may be unrepresented.

A further challenge for effective treatment is identification of the causative agent. Calf diarrhoea and BRD are not caused by single pathogens but are generally composite syndromes where clinical signs depend on pathogens, host and environment and their interactions.16 Consequently, focusing on clinical state rather than the putative pathogen can be an effective strategy to identify disease problems in both clinical and research contexts.1 14 17 Many aspects of calf housing also impact on calf health. These include building design (ventilation, temperature, humidity, etc), cleanliness of the pens and feeders, pen size and proximity to other calves and older cows.18 19 Receiving an adequate supply of good quality colostrum within the first 12 hours after birth is considered essential for calf health by providing passive immunity.20 After ingestion, IgG is transferred to the calf’s circulation, whereas IgA acts locally within the intestinal mucosa.21 Calves receiving inadequate colostrum experience a failure of passive transfer (FPT), which is assessed by measuring circulating levels of IgG and/or total protein (TP). Values less than 10 mg/ml IgG or less than 50 mg/ml TP have generally been defined as the cut-off points for FPT, although others have used 12 and 55 or 57 mg/ml, respectively.22–26 Given that contagious diseases affecting calves have a multifactorial nature, they are influenced by a dynamic and varied group of factors. However, accurate health information about dairy heifers on the UK commercial farms is currently lacking, with recording of calf disease and mortality being much lower than in Scandinavian countries.27 Youngstock are rarely included in computer-based recording systems used for the adult herd and paper-based records are often incomplete or inaccurate. Only 39 per cent of Welsh farmers in a recent survey recorded calf disease, with veterinarians describing records as adequate on only 8 per cent of farms for diarrhoea and 16 per cent for respiratory disease.28 Farmers are consequently often unaware of reliable disease or death rates for their own calves. The UK is a significant dairy-producing country with around 1.9 million adult dairy cows29 and consequently this reflects a significant gap in existing knowledge.

The aim of this prospective cohort study was to monitor preweaned heifers on commercial farms with weekly health visits to assess their health and passive transfer status and to determine whether areas of the UK dairy heifer management require improvement. The evidence on farmer recording cited above suggests that trained technicians or veterinarians are required to ascertain disease rates reliably. This is the first in a series of four papers and provides a description of the cohort, key management practices, measures of passive transfer and disease incidence. Future papers will address the factors associated with disease risk in individual calves, calf growth rates and aspects of colostrum management.

Materials and methods

Cohort design and recruitment

A convenience sample of 11 dairy herds in S.E. England were recruited by approaching local veterinary surgeries to suggest farmers to participate. Farms were then visited to discuss the project and complete a consent form. For inclusion in the study, farms needed to expect 30–50 heifers to be born during a recruitment period of four months and ranged in size from 186 to 550 adult cows. There were two calf recruitment periods, either October 2011 to February 2012 (seven farms) or August to October 2012 (four farms). During these times, all births were recorded and all heifer calves alive after 24 hours were recruited at weekly visits. A literature review revealed that while little was known of calf disease rates in the UK, the incidences of diarrhoea and BRD were usually between 10–25 per cent and 7–29 per cent, respectively in other countries which have included the USA, Canada, Holland and Sweden.4 Assuming a rate for each of these diseases of around 20 per cent, a target cohort of 500 was considered adequate to allow for subsequent comparisons between sick and healthy animals.

Herd profiles including management practices

Information on herd details and calf rearing practices for each farm were collected. This initially involved an interview with the farmer or herdsmen using a standardised data capture template as described by Brickell and others.30 These farm records were subsequently updated and additional information on the individual calves was obtained during the weekly data collection visits. Information on the farm included herd size (number of adult cows, number of heifers being reared), calving pattern, predominant breeds kept, ownership and staffing and average milk yield. Calving information included where the cows calved (group paddock, straw yard, individual pen) and age of separation of calf from dam (snatched, 12–24 hours, or left with dam for >24 hours). For each live heifer calf recruited onto the study, data were obtained on whether dystocia occurred on a 3-point scale (1, no assistance; 2, mild pull; 3, traction). Routines for calf housing on each farm and group sizes were recorded and it was noted if the calves were transferred to another pen before weaning. If so, the age at which this occurred and whether or not it involved mixing calves from different initial groups was recorded. Details of vaccination policy for preweaned calves were obtained. Water availability was noted at each visit, including issues with spilling and fouling that prevented adequate access to palatable water.

Measurement of passive transfer

Farms recorded which calves received supplementary colostrum by bottle or tube. A jugular blood sample was taken into a plain Vacutainer (BD, Oxford, UK) at the first clinical examination, when calves were between one and eight days old. These samples were collected under the UK Animals (Scientific Procedures) Act 1986,9 under project licence PPL 70/7223, also approved by the RVC Ethical Review Process. After clotting, the sample was centrifuged and the serum stored at −18o C for later analysis. A refractometer (RHC-200, Linx, China) was calibrated with distilled water then used to assess serum TP. IgG was measured using a radial immunodiffusion assay kit (VMRD, Pullman, Washington, USA) according to the manufacturer’s instructions. After 18–24 hours incubation at room temperature, each gel was scanned using a gel reader (G:Box, Syngene, Cambridge, UK) and the diameter of each precipitate ring was then measured using ImageJ (

Collection of health data

Farmers were asked to record details of calving and calf disease on preprinted white boards and this information was checked at each weekly visit. Calving information was well recorded but disease information was not added consistently so could not be included in the analysis. Where farmers were already treating an animal, their presumed diagnosis was recorded and confirmed at the next clinical examination. The clinical scoring system for calf health was developed by the University of Wisconsin-Madison and used a systematic approach to assess faecal consistency and signs of BRD.1 Some minor modifications were made for use in the UK as summarised in Table 1. Several areas were adopted with no changes: ocular and nasal discharges and presence of induced or spontaneous cough were scored on a scale from 0 to 3. Head tilt, a sign typical of inner ear infection, was excluded because it was not seen in any calf in the study. Temperature scoring used cut -off values of 38.5oC, 39oC and 39.5oC (rather than 101oF, 102oF and 103oF). Animals with a respiratory score ≥5 were classified as having BRD. Faecal scoring was used as described in Table 1 and those calves with a faecal score greater than and equal to 2 were classified as having diarrhoea. Notes were also taken at each visit to record any other problems, most commonly umbilical infection. Clinical examinations were completed on every calf each week for nine consecutive weeks by the same research veterinarian.


Health scoring system used in the weekly clinical examinations, adapted from McGuirk1 for use in the UK

For ethical reasons, both the farmers and their local veterinary surgeon were informed after each visit of any animals showing clinical signs of disease. Each farm had agreed treatment protocols in place. For BRD, this always included a licensed antibiotic and a non-steroidal anti-inflammatory drug. In the UK, these must legally be recorded in a medicines book; these books were checked weekly to determine which calves had been treated for BRD. No farmers kept any calf disease records other than their medicines books. For diarrhoea, the treatment protocol used rehydration solutions which did not legally have to be recorded. In practice it was not, therefore, possible to obtain exact information about which calves were fed rehydration solution or how many times they were treated. Where farmers described having treated calves but there were no signs of any disease on clinical examination, the clinical signs recorded by the farmer were used. All calf deaths were recorded and the presumed cause was established by the local veterinarian by means of a gross pathology postmortem examination.

Data analysis

Data were stored in Excel and arranged into a relational database using Access to allow for further analysis (Microsoft Office, Microsoft, Redmond, Washington, USA). Missing data were stored as an ‘NA’ for not available and no interpolation was used. All analysis was completed in R using the lattice package for graphics ( The passive transfer measurements were normally distributed, so these results are presented as mean (±sd). Other data on mortality, disease and colostrum management were count data and results are presented as the count and the percentage. To calculate incidence rates for disease, the number of new cases was divided by calf weeks at risk. New cases were defined as a calf meeting the diagnostic criteria that had been scored as healthy for that condition in the previous week. For incidence rate calculations, this meant that repeated cases were included but continuing disease was not. The calculations on how many weeks a calf had a disease used the total number of positive clinical examinations and did not account for difference between long disease duration compared with repeated cases. CIs to compare proportions of animals affected between farms were calculated using the test of equal proportions (prop.test). Differences between farms were assessed by analysis of variance (ANOVA) with post hoc testing using the Tukey HSD method in the R package ‘agricolae’ to correct for multiple testing. A Student’s t test was used where there were only two groups for comparison. The level of significance was set at P<0.05.


Herds and breeds

Three of the 11 farms (A, H, K, Table 2) had an autumn block calving pattern (September to November), with the remainder having all-year-round (AYR) calving. Five farms were pure Holstein, three were Holstein with some additional Viking Red crosses, and one farm had Holsteins alongside Ayrshires. Of two farms using more grass-based systems, one kept predominantly New Zealand Friesian type animals along with some Viking Red cross Friesian and the other had a large number of mixed breed cattle with many Jersey and Jersey crosses alongside Friesian cross heifers. From these farms, 492 dairy heifer calves were recruited, with a range of 26–56 calves per farm (Table 2). In total, 351/492 (71  per cent) of these calves were Holsteins with the next largest groups being black and white cross Viking Red cattle (43/492, 8.7  per cent) and Friesians (37/492, 7.5  per cent). Of the 470 calves that survived until the nine-week clinical examination, 403 had a complete data set (85.7  per cent), 45 had a single missing data point (9.8  per cent) and 22 calves (4.5  per cent) had more than one week of missing data.


Summary of calf management practices for the 11 farms included in the study

Preweaning feeding and housing management

The management practices for calving and calf management are summarised in Table 2. The three autumn calving herds calved outside in group paddocks. Six herds calved in groups in straw yards. Two farms used individual calving pens, with one of those snatching calves soon after birth as a Johne’s disease control strategy. Most other farms left calves with their dams for an estimated 12–24 hours, although this was highly variable and some individual calves on one farm remained with their dams for up to 10 days.

The initial group size ranged from single pens up to groups of 20 calves. Four farms fed calves using automated feeders. Three of these calves were moved from the farm of origin to a nearby rearing unit once passports were issued (typically at one week old). The other kept calves at the farm of origin and bucket fed twice daily for approximately three weeks until a batch was moved to a rearing unit. Three farms fed milk twice daily, two using buckets and keeping calves in single pens and two using feeders with teats and keeping calves in groups of 12. A further two farms fed calves twice daily for approximately two weeks and then switched to once-daily milk feeding. This study was completed in 2012, before the clarification issued by the Department for Environment, Food and Rural Affairs (DEFRA) that once-daily feeding under 28 days of age was impermissible.31 Finally, two farms kept calves in small groups where there was a single teat per pen of four to six calves supplied by a mixed drum of acidified waste milk. All farms provided forage and commercial calf pellets fed ad libitum from the first week of life until after weaning. All farms bedded on deep straw: on three farms this was the only source of fibre, four farms fed additional chopped straw off the ground, two fed hay and two offered silage to calves. No farmers provided water in the pens used to house calves for the first few days. All farms then provided water, although there were regular issues with spilling and fouling that often prevented adequate access to palatable water. Farm F did not provide water premovement to their rearing unit, which occurred at an average age of 29 days.

All farms used gradual step-down weaning, typically over 1 week beginning at 6–10 weeks of age. Most farms based their timing decision on a combination of calf age, calf or group appearance and convenience. The interquartile range for weaning age was quite narrow, with half of all calves weaned between 60 and 69 days of age, but the overall range from 34 to 96 days was much wider (Fig 1).

FIG 1:

Histogram illustrating range of weaning ages for the 492 calves in the study.

Colostrum feeding management and passive transfer

No farms monitored either colostrum quality or passive transfer outside the study. The supplemental colostrum used was previously frozen from the first milking and generally fed at 2×2 litres. No farms used artificial colostrum supplements. Farm F pasteurised colostrum as a Johne’s disease control measure. The two largest AYR calving farms routinely fed supplemental colostrum using either an oesophageal tube or bottle (Table 3). Two further farms gave supplementary colostrum to most calves (75 per cent and 84 per cent, respectively). The remaining seven farms supplemented when staff felt it was required, but on these farms only between 0 per cent and 19 per cent of calves received any additional colostrum during the study period. In total, 35.2 per cent (173/492) of calves received supplementary colostrum.


Details of passive transfer by farm as measured in the first week of life*

The mean serum IgG in the first week of life was more than 10 mg/ml on all farms, indicating adequate passive transfer (APT), with an overall mean of 19.0±10 mg/ml IgG (Table 3). The variation within farms is illustrated in Fig 2. The three farms with autumn block calving (A, H, K) had high mean IgG and a low number of calves with FPT, although they only supplemented calves rarely. Three farms had more than 25 per cent calves with less than10 mg/ml IgG. The overall mean value for TP was 56.7±10.3 mg/ml. One farm had an average TP of 48.8±8.5 mg/ml, below the accepted threshold for APT of more than 50 mg/ml. Overall, 20.7% of calves were recorded as having FPT when classified using IgG less than 10 mg/ml and 24.4 per cent using TP less than 50 mg/ml. As described in the ‘Introduction’ section, a range of cut-off points has been recommended previously. There was a large difference between the two methods in classifying calves with marginal results, with 4.1 per cent of calves having an IgG of 10–12 mg/ml but 21 per cent having a TP of 50–55 mg/ml.

FIG 2:

Box and whisker plots showing (A) serum IgG measured by radial immunodiffusion and (B) serum total protein measured by refractometer. Samples were taken at calf recruitment, from 1 to 8 days after birth. Results are split by farm of origin and farms are ordered by median value. Horizontal lines indicate the commonly used reference values: in (A) red is at 10 mg/ml and green at 12 mg/ml and in (B) red is at 50 mg/ml and green at 55 mg/ml. Letters show the results from analysis of variance: groups with the same letter did not differ significantly (P>0.05).

Calf mortality

For the cohort as a whole, 3.1 per cent of calves died in the first month and 1.4 per cent in the second, giving an overall mortality rate of 22/492 (4.5 per cent) (Table 4). The main causes of death were diarrhoea (n=8, all in the first month) or BRD (n=7, three in the first and four in the second month). Other causes were dystocia-related (n=2), omphalophlebitis (n=1), neonatal pancytopenia associated with BVD vaccination of the dam (n=1), rumen bloat (n=2) and accident (n=1). All farms had at least one death among their preweaned calves, ranging from 1 to 6. Of the total diagnosed cases for diarrhoea and BRD, 2.6 per cent (8/308) and 1.8 per cent (7/396) died, respectively.


Duration and frequency of diarrhoea and bovine respiratory disease in preweaned calves based on once-weekly assessment using a modified Wisconsin-Madison health scoring system*

Treatment of preweaning disease

Calves showing signs of clinical disease at each weekly examination were reported to the farmer and their veterinary surgeon and treated according to an established protocol as described above. It was not therefore possible to assess any treatment effects since all sick calves were treated. Farm I used a routine treatment with a long-acting macrolide antibiotic, for example, tilmicosin (Micotil, Elanco, Basingstoke, Hampshire, UK) or tulathromycin (Draxxin, Zoetis, London, UK) when calves were moved at 10 days old, in an attempt to reduce BRD. Veterinary surgeons were called out to treat animals at six farms on 10 occasions, three for group problems (one diarrhoea, Farm H) and two BRD outbreaks (Farms I, K). Five individual animals were treated by their veterinary surgeon for diarrhoea (n=2), BRD, neonatal pancytopenia and an umbilical injury.

Incidence of diarrhoea

Throughout the whole preweaning period, 48.2 per cent of calves were diagnosed with diarrhoea with an overall incidence rate of 7.8 cases/100 calf weeks at risk (95 per cent CI: 7.1 to 8.8 per cent). The highest proportion of examinations where a calf showed signs of diarrhoea was at the second clinical examination, when 126 calves (28.2%) reached the threshold for diagnosis (Fig 3A). Very few diarrhoea cases were observed in calves over 4 weeks old, with less than 3% of older individuals affected in any week. There was a wide range of disease rates from 24.1% to 74.4% on different farms (Table 4). While the highest incidence (74.4 per cent) was on one of the larger farms (F), there was no overall relationship with herd size (ANOVA, P=0.26). Most calves showed clinical signs for one week, with only 24.0 per cent of calves affected for more than one week. On Farm F, however, 11/29 (37.9 per cent) of the calves with diarrhoea showed clinical signs for greater than and equal to2 weeks. Faecal scoring was used as the only method of diagnosing diarrhoea, but calves with diarrhoea had a higher mean temperature than those with normal faeces (healthy, 38.8oC (95  per cent CI: 38.78 to 38.89) vs diarrhoea, 39.01oC (95 per cent CI: 38.95 to 39.07), P<0.001). Overall, calves had a temperature more than 39.0oC (defined as pyrexia by McGuirk1 on 177/307 (57.6  per cent) of occasions where diarrhoea was detected.

FIG 3:

Number of clinical examinations performed on all 492 calves between recruitment and nine weeks of age. The green bars represent the number of examinations where calves were unaffected by disease and the red bars show the number of calves showing signs of (A) diarrhoea and (B) bovine respiratory disease based on assessment using a modified Wisconsin Madison calf health scoring system.

Incidence of BRD

BRD showed a different pattern, with cases spread throughout the preweaning period (Fig 3B). The incidence rate was 10.1 cases/100 calf weeks at risk (95per cent CI: 9.2 to 11.2 per cent). The smallest number of cases was observed in the first week, when 5.1 per cent of clinical examinations were positive. Cases of disease were high throughout the second month with more than 10 per cent of individuals affected in weeks four to nine. BRD of any duration was diagnosed in 45.9 per cent of calves (95 per cent CI: 41.2 to 50.3 per cent). There was a large variation in the proportion of calves affected between farms, ranging from 20.4 to 77.8 per cent (Table 4) and this was not related to herd size (ANOVA, P=0.14). Four farms had more than 60 per cent of calves showing BRD preweaning. Of those calves affected, 43.4 per cent showed clinical signs for more than one week. Farm J, with the least number of calves affected, also had the shortest duration of disease with 10/11 (90.9 per cent) individuals only showing signs of BRD for a single clinical examination. In contrast, only 28.2 per cent of calves on Farm F were affected but these had the longest duration of disease with 7/11 (63.6 per cent) showing clinical signs for two or more weeks. In calves diagnosed with BRD, 83.3 per cent had a rectal temperature greater than and equal to 39°C. The mean rectal temperatures were significantly different (P<0.001): healthy, 38.8oC (95 per cent CI: 38.73 to 38.83) vs BRD, 39.4oC (95 per cent CI: 38.34 to 39.44).


This study has quantified passive transfer, health and mortality in preweaned dairy heifers on 11 commercial UK dairy farms. These were all typical commercial farms, although they were relatively large for the UK, ranging in size from 186 to 550 adult cows, compared with the UK average of 133.29  Calf management was similar to that seen in a stratified sample of 102 UK dairy herds.8 This was rarely a specialised task and generally involved several farm staff. There were few centralised methods of data recording for this age group and, as has been shown in other studies,28 there was little recording of disease in heifer calves by farmers. Calf health has been considered here at a farm level rather than an individual calf level; risk factors for disease in individual calves will form the basis of the next paper. In practice, calves are managed as groups with farmers and their veterinary surgeons making most herd health and management decisions based on group performance.

Disease cases are affected by a triad of factors and their interactions: the environment, the pathogen and the calf. When considering a cohort born on the same farm at a similar time of year, two of these three factors remain reasonably constant. Disease was assessed using an adapted Wisconsin-Madison Scoring system, which is now widely established as a useful systematic way to determine calf health.1 Clinical scoring could be undertaken easily and quickly by one person and standardised definitions of disease allow comparison with other studies. This method was not, however, a good tool to discriminate between upper respiratory tract infections and pneumonia. Additionally, for both BRD and diarrhoea, the weekly assessment interval used means that some cases recorded as continuing cases could potentially have been relapses. The influence of disease treatments could not be investigated here as all sick animals were treated promptly according to standardised protocols.

The incidence of mortality among heifers which were born alive but died in their first month was similar to that reported in a previous UK cohort, at 3.1 per cent compared with 3.4 per cent2; mortality in calves in the second month of life was lower at 1.4 per cent. Diarrhoea and BRD each accounted for about one-third of the deaths. All incidences of disease were reported to the farm and the veterinarian on a weekly basis, with comparative overall results between farms reported both midway through and at the end of the study for benchmarking purposes. The weekly reports might potentially have improved diagnosis and treatment and so reduced the fatality rate.

A previous UK study found that 26 per cent of 444 calves tested had FPT (range 5–51 per cent between seven farms tested,32 compared with 20.7 per cent in this cohort. In comparison, a much higher proportion of 48 per cent of 414 Welsh calves sampled had FPT based on zinc sulphate turbidity testing (ZST).28 Beam and others33 undertook a large-scale assessment of colostrum management in the USA and found that 19.2 per cent of calves experienced FPT even though 76 per cent were reported to receive supplementary colostrum. In the present study, there was a much lower intervention rate of 35.2 per cent with a mean IgG of 19 mg/ml. There was, however, high variability between farms for mean IgG and TP results. Farms with delayed cow/calf separation and autumn block calving had better passive transfer than farms that separated calves immediately and had AYR calving patterns (Table 3). This suggests a possible conflict between Johne’s disease management, where disease spread can be reduced by early separation and feeding heat-treated colostrum,34 against the good passive transfer that appears to occur when cow and calf are uninterrupted. More tailored advice may therefore be useful for farmers, depending on their Johne’s disease status.

Despite the good passive transfer, contagious disease was highly prevalent, with 48 per cent of calves diagnosed with diarrhoea and 46 per cent with BRD. The farms included in the study were not a random sample. It is therefore possible there was bias in farm selection, although no disease-related criteria were used for inclusion. Another recent UK cross-sectional study also found a high disease incidence, but only 39 per cent of farmers were recording calf disease, meaning that high rates may be going unnoticed.28 Diarrhoea incidence was higher than that reported in most other cohorts on commercial farms in other countries, which ranged from 10 to 35 per cent (reviewed by Johnson and others4). One study with a higher incidence of 64 per cent deliberately recruited farms with a known diarrhoea problem: diagnosis was based initially on clinical signs and samples were then taken for pathogen identification.35 In the present study, diarrhoea diagnosis was only based on clinical signs so it is possible that some of the cases were non-contagious diet-related scours. Diarrhoea can also be associated with rumen drinking in bucket-fed calves.36 Milk may enter the rumen due to failure of closure of the oesophageal groove causing acidosis and inflammation and potentially leading to rumen ulceration and secondary bacterial and fungal infection. However, samples from 91.2 per cent of diarrhoeic calves in another study tested positive for enteric pathogens.37 Another limitation of the current study was that all the calves recruited were born between August and February. This coincided with peak calving periods but did mean that spring and summer births were omitted. Season is assumed to be an important factor in calf disease, although there is conflicting evidence as to when higher disease and mortality rates may be expected.11 38 Some studies on research farms have included all calves born over several years.13 It is rare that this can be achieved on commercial farm, although future work should address this question of seasonality in the UK.

BRD was more common than in any recently reported cohort, which reported incidences ranging from 1 to 39 per cent.4  Comparing two separate studies in New York State, USA, the reported incidence of BRD was 21.6 per cent higher when it was recorded by veterinarians10 compared with farmers.39  The high incidence found here may reflect the relatively intensive monitoring regime used in the current study, for example, compared with less frequent visits40 or with farmer reporting.14 Disease was usually first diagnosed at the clinical examination rather than by farmers. This is to be expected as farmers often use reduced milk consumption as the main sign of ill health and, when calves are fed limited rations, this is a late sign of disease.27 The diagnostic criteria used for both diarrhoea and BRD were nevertheless similar to several previous studies, suggesting that the disease rates on these UK farms really were comparatively high.

Of calves with BRD, almost half had clinical signs for longer than one week including pyrexia. This is associated with changes in behaviour and poor welfare through the expression of sickness behaviours41 suggesting that the high rates of BRD recorded indicate potential problems for calf welfare. Once contagious disease has occurred, there is increased risk of future morbidity: previous diarrhoea was associated with repeated loose faeces42 43 and later BRD.39 42 44

Diarrhoea in the first month was the single most frequent cause of death in this cohort, despite good passive transfer. Recent literature has suggested that there is commonly no significant relationship between contracting diarrhoea and passive transfer measured in blood. For example, only 10.3 per cent of calves had a serum TP less than 52 mg/ml in a recent Canadian study on farms with an existing diarrhoea problem.35 Similar large cohort studies of 200–600 heifers also found no correlation.10 14 17 35 In another recent cohort of 2874 heifers recruited from 19 commercial dairy farms, failure of passive transfer was not associated with neonatal diarrhoea when other factors were accounted for in the analysis.26 A complexity of investigating this type of relationship is that the pathological effects of diarrhoea can cause either dehydration or protein loss via the gut, so TP may be abnormally high or low in diarrhoeic calves. In the present study, serum IgG was recorded in addition to TP in an attempt to mitigate this problem. Increasing use of serum total protein to monitor passive transfer24 35 supports the need for further study on an individual calf basis to explore this relationship.

There was regularly insufficient availability of water on some farms despite a legal requirement to provide water to calves aged over one week. This generally occurred when buckets were spilled or fouled, but was also observed when calves were late in transfer to a rearing unit. In a previous survey of 102 UK dairy farms, only 78 per cent of farms provided water to calves within seven days of birth and two farms did not provide water until after weaning.8 It was noticeable that diarrhoea typically occurred in very young calves which generally had little access to water. A lack of water, especially if associated with a restricted milk ration, would worsen dehydration and contribute to mortality. Most calf diarrhoea is contagious in origin.35 37 Therefore, to prevent diarrhoea, farmers should be advised to focus on improving hygiene of housing and feeding. Although cleanliness has excellent evidence to support its benefits,4 19 this strategy was recently reported to be a low priority among Welsh farmers.28

Analysis of results on disease incidence with respect to herd size have varied, and no associations were observed in this study. In previous studies, the risks of more calf mortality3 45 Cryptosporidium  parvum shedding46 and the number of cases of BRD diagnosed44 were all higher in larger herds. However, in an extensive study of around 47 000 heifers in the USA, farm size showed no association with mortality and lower producing units (which tended to be smaller) had higher mortality.47 It is possible that herd size can interact with other factors explaining the associations found in some previous reports. This will be examined in the next paper on this cohort to determine the main risk factors for disease on an individual calf basis.


In summary, this cohort of UK heifers performed fairly well in some areas: passive transfer was better than or similar to other cohorts of heifers studied, with only 21 per cent of animals having IgG less than 10 mg/ml. The mortality rate of 4.5 per cent was also relatively low. This cohort nevertheless performed worse than those studied internationally with respect to the incidence of contagious endemic disease with very high rates of both BRD and diarrhoea. Although this did not appear to cause high mortality, disease does indicate poor calf welfare. There were also clear needs for improved provision of clean water for any sick animal and all calves over seven days old, to meet the legal requirement. In order to minimise disease in their stock, farmers need additional training in prompt recognition and recording the signs of disease and they should also be encouraged to monitor their mortality rates. They can then work with their veterinary surgeons to use this knowledge to optimise their general management strategies and to develop protocols to identify and respond to disease appropriately.


  1. 1.
  2. 2.
  3. 3.
  4. 4.
  5. 5.
  6. 6.
  7. 7.
  8. 8.
  9. 9.
  10. 10.
  11. 11.
  12. 12.
  13. 13.
  14. 14.
  15. 15.
  16. 16.
  17. 17.
  18. 18.
  19. 19.
  20. 20.
  21. 21.
  22. 22.
  23. 23.
  24. 24.
  25. 25.
  26. 26.
  27. 27.
  28. 28.
  29. 29.
  30. 30.
  31. 31.
  32. 32.
  33. 33.
  34. 34.
  35. 35.
  36. 36.
  37. 37.
  38. 38.
  39. 39.
  40. 40.
  41. 41.
  42. 42.
  43. 43.
  44. 44.
  45. 45.
  46. 46.
  47. 47.
View Abstract


  • Contributors All authors contributed to the design of the study. KFJ and NC collected the samples and KFJ analysed the data. DCW wrote the first draft of the paper. All authors read and approved the final manuscript.

  • Funding The study was funded by BBSRC (Grant Reference: BB/FO16891/1) and Volac International Ltd. and made possible by all the farmers and their veterinary surgeons who participated.

  • Competing interests None declared.

  • Provenance and peer review Not commissioned; externally peer reviewed.

  • Data sharing statement No additional data are available.

Request Permissions

If you wish to reuse any or all of this article please use the link below which will take you to the Copyright Clearance Center’s RightsLink service. You will be able to get a quick price and instant permission to reuse the content in many different ways.