We will have two real experts on heat pumps addressing our afternoon conference “Decarbonising heat – practical realities” on 8 July, which will focus on the non-domestic market and the lessons that can be learned from real-life installations.
Ben Whittle is a technical manager from the Energy Saving Trust. He has been working in the world of renewables for 20 years, and has previously worked for companies designing and installing solar thermal, solar PV, biomass and heat pump systems to megawatt scale.
John Cantor started out manufacturing and installing bespoke heat-pump systems during 1980s and 90s. He was system inspector for the first UK grants through BRE and was also on the MCS working group. He’s an honorary member of the Ground Source Heat Pump Association and author of ‘Heat Pumps for the Home’.
As well as heat pumps we’ll examine the realities of biomass installations and consider the prospects for hydrogen. More details at https://vesma.com/z200
Delighted to have Jan Rosenow opening our afternoon conference “Decarbonising heating – practical realities” on 8 July. Focussing on the non-domestic market, our speakers will discuss real-life experience with biomass boilers and heat-pump systems, and draw lessons for future projects. We’ll also hear from an expert on hydrogen about how that might be introduced into the national heating-fuel mix, and we have allowed plenty of time for questions from the audience.
WHEN thinking about possible energy-saving projects, you might ask yourself how radical you want them to be. There’s a spectrum from the relatively easy to the costly and disruptive. Readers may have their own views on this but I think the spectrum runs like this:
Make sure control setpoints and timings are correct. This is generally the cheapest and least disruptive measure one can take;
Enhance the control strategies. For example introduce floating setpoints on chilled water circuits, optimum start in place of fixed timers, or variable-speed control of motor-driven equipment;
Implement loss-reducing modifications. Examples here include zone isolation valves in a compressed-air network or thermal insulation on hot pipework;
Improve component efficiency. The classic case here is lighting technology (which may be cost-effective in its own right) but think also about things such as the introduction of higher-efficiency or better-sized motors (which may only be economical when replacement is necessary for other reasons);
Improve process layouts and integration. Here I am thinking primarily about opportunities for waste heat recovery, but there are other special cases where part-processed materials may gain or lose moisture or heat while in transit between stages to the detriment of overall energy efficiency. And finally the nuclear option:
Retire buildings or process plant in favour of more energy efficient replacements
Effective energy waste avoidance relies crucially on the comparison of actual and ‘expected’ consumptions. Classically we do this on a weekly or monthly basis, using models for expected consumption that are linked to independent driving factors. But there are other ways to skin that cat.
Buildings will in many cases have a characteristic diurnal pattern of demand that can be expressed as a profile at, say, half-hourly intervals. With a large enough group of similar buildings, and taking account of drivers like the weather, it seems possible in theory to create a dynamic template for each building against which its demand can be assessed in near-real-time. The template is just a different way of calculating and expressing expected consumption, but it creates the realistic prospect of daily exception reports. Of course the implied excess costs need to be taken into account, because you need to be able to suppress the clutter of insignificant deviations, prioritise cases for investigation and estimate the value of resolving them, just as you would if you were using a weekly or monthly overspend league table.
The role of artificial intelligence here is to learn what ‘correct’ behaviour looks like and one advantage of this in large estates is that it obviates the need for human analysts to calibrate degree-day regression models for every meter. Another benefit would be the recognition of common abnormalities in profiles. Properly trained with correct human feedback, an AI-based pattern recognition system could in principle recognise symptoms that have occurred before elsewhere and associate them with remedies that have previously been successfully applied.
A further benefit is advanced benchmarking. In classical M&T we know that buildings can be benchmarked by comparing the slopes and intercepts respectively of their degree-day regression lines. A pattern-analysis system can take this more incisive analysis to a whole new level.
I will be interviewing James Ferguson, a keen proponent of AI in energy waste detection, on 15 July 2021 in my “Energy Conversations” series of open video calls. If this is a subject which interests you you can request a place in the audience here.
Once you have discovered how to routinely calculate expected consumptions for comparison with actual recorded values, you can get some very useful insights into the energy behaviour of the processes, buildings and vehicles under your supervision. One thing you can do is chart the history of how actual and expected consumption compare. In this example we are looking at the daily electricity consumption of a large air-compressor installation:
The green trace represents expected kWh (computed by a formula based on the daily air output) and the individual points represent the actual metered kWh. Most of the time the two agree, but there were times in this case when they diverged.
It is illuminating to concentrate on the extent to which actual consumption has deviated from expected values, so in the following chart we focus on the difference between them:
There will always be some discrepancy between actual and expected consumptions. Part of the difference is purely random, and the limits of this typical background variation are signified by the red dotted lines. If the difference goes outside these bounds, it is probably because of an underlying shift in how the object is performing. In the above diagram there were three episodes (one moderate, two more severe) of abnormal performance. Significant positive deviations (above the upper control limit) are more usual than negative ones because consuming more energy than required for a given output is much more likely than using less.
For training in energy consumption analysis look for ‘monitoring and targeting’ at VESMA.COM
In a well-constructed energy monitoring and targeting scheme, every stream of consumption that has a formula for expected consumption will also have its own control limit. The limits will be narrow where data are reliable, the formula is appropriate, and the monitored object operates in a predictable way. The limits will be wider where it is harder to model expected consumption accurately, and where there is uncertainty in the measurements of consumption or driving factors. However, it is not burdensome to derive specific control limits for every individual consumption stream because there are reliable statistical methods which can largely automate the process.
Control charts are useful as part of an energy awareness-raising programme. It is easy for people to understand that the trace should normally fall between the control limits, and that will be true regardless of the complexity of the underlying calculations. If people see it deviate above the upper limit, they know some energy waste or losses have occurred; so will the person responsible, and he or she will know that everyone else could be aware of it as well. This creates some incentive to resolve the issue, and once it has been sorted out everyone will see the trace come back between the limits.
Widespread adoption of automatic meter reading has given many energy users a huge volume of fine-grained data about energy consumption. How best to use it? A ‘heat-map’ chart is a powerful visualisation technique that can easily show ten weeks’ half-hourly data in a single screen. This for example is the pattern of a building’s gas consumption between November and January:
Each vertical slice of the chart is one day, running midnight to midnight top to bottom, with each half-hourly cell colour-coded according to demand . This creates a contour-map effect and when you look at this specifi example, you can see numerous features:
Fixed ‘off’ time;
Optimised startup time (starts later when the building has not cooled down as much overnight);
Peak output during startup;
Off at weekends but with some heating early on Saturday mornings;
Shut-down over Christmas and New Year; and
A brief burst of consumption during the Christmas break, presumably frost protection.
Further examples
This building’s gas consumption pattern is quite similar to the previous one’s (they both belong to the same organisation), but the early-morning startup boost is much more evident and occurs even during the Christmas and New Year break:
Next we have a fairly typical profile for electricity consumption in an office building. What is slightly questionable is the higher daytime consumption near the start (April) compared with the end (June). This suggests the use of portable heaters. Note also that the peak half-hourly demands can easily be seen (Friday of the second week and Wednesday of the fiifth week). In both cases it is evident that those peaks occurred not because of any specific incident but because consumption had generally been higher than usual all day:
In this final example we are looking at short-term heatmap views of electricity feeding a set of independent batch processes in a pharmaceutical plant. The left-hand diagram is the actual measured consumption while the right-hand diagram is the expected profile based on a mathematical model of the plant into which we had put information about machine scheduling:
This case history shows how fine-grained energy measurements enabled savings to be verified in the face of assorted practical problems.
THE STORY concerns a large air-conditioned establishment in the middle east which had been fitted with adiabatic cooling sprays on its central chillers. It is one of a number belonging to an international chain and my purpose was to establish whether similar technology should be contemplated elsewhere in the chain.
I had originally been commissioned to give a second opinion on the savings claims made by the equipment supplier. Although the claims were quite plausible, they lacked some credibility because they were based on an extremely short evaluation. So in an initial effort at independent checking I obtained several years’ worth of monthly consumption data at the whole-site level and analysed it against local cooling degree days. The results were ambiguous because there appeared to be some unrelated phenomenon at play which resulted in the site toggling on a seasonal cycle between two performance characteristics, one significantly worse than the other, with an impact sufficient to mask the beneficial effect of the energy conservation measure (ECM).
Without reliable evidence I declined to verify the supplier’s assessment, and recommended a deliberate test using a ‘retrofit isolation’ approach based on three existing electricity submeters (one per chiller) and a new heat meter in the common chilled-water circuit. Because nobody wanted to pay for a proper heat meter, a clamp-on ultrasonic flow element was specified. At this point the pandemic struck.
Because we had data-logged metering, and because the adiabatic cooling system sounded as if it would lend itself to being turned on and off, the measurement and verification plan was based on day-on, day-off testing. I had proposed a ten-week testing campaign which would give us 35 observations in each state. The plan called for two regression models to be developed: one for the ECM ‘on’ days and the other for the ‘off’ days. Extrapolation from the regression formulae would indicate a percentage difference and variance within each would confirm how much uncertainty there was.
Then came the collision with reality. The ECM supplier said disabling it wasn’t as simple as turning the cooling sprays off; he insisted that the mesh screens which were part of the installation should come off as well and be reinstated when the spray was re-enabled. This was a quite reasonable stance because it gave a more valid comparison, but unfortunately he couldn’t afford to send a man to site every day to do what was necessary. Meanwhile the establishment’s manager had got wind of the project and put pressure on the site engineer not to disable the ECM, which he was convinced was saving him a lot of money which (thanks to lockdown) the business could not afford to lose. He also had a point. Luckily we agreed a compromise: in exchange for them coming down to a three-day on-off cycle, I promised to monitor things closely and terminate the test as soon as conclusive results emerged.
Needless to say, the ultrasonic meter let us down. The technician responsible for downloading its data reported that some of its hourly totals looked suspiciously high, and he proposed filtering them out. But I feared that he might not be able to capture all the rogue points. Some of them might only be slightly wrong, but wrong nonetheless. When we drilled down we discovered that the raw data were stored at one-minute intervals, with the measurement fault manifesting itself as gross errors confined to occasional one-minute records. You can see this from this figure, which spans several months but plots every single one-minute record:
Valid readings at the one-minute interval clearly fell below a threshold of about 450 m3/hr, and abnormal readings were (a) very clear and (b) sparsely distributed through time, so my colleague Daniel Curtis was able to pushed the records through a sieve, take out the lumps and thereby cleanse the data (and that included reinstating hourly values which the meter software had censored because they appeared to be too low). We were helped in this by the flow normally being relatively constant, so that simple interpolation was an accurate gap-filling strategy. When we compared reported flow measurements with corrected values at eight-hour intervals we saw that in reality almost every reported value had been wrong:
Analysis then proceeded using cleaned-up eight-hour interval data, but we were still not out of the woods. From the individual chillers’ electricity meters we could see that the site operations team were changing the chiller sequence from time to time. Fortunately their manager agreed that barring emergencies they could leave the same chillers in service, with the same one in the lead, until the test finished. They also agreed not to change the chilled-water set point temperature, which we discovered they were in the habit of tweaking. Of course in a textbook measurement and verification exercise these static factors would have emerged during planning but this project was being conducted in a hurry on a shoestring and managed remotely. That wasn’t the end of it: later in the test we would have interruptions because the chiller maintenance firm was scheduled to clean the condenser coils. More on that later.
Nor were those the only complications. A few weeks after we started testing in anger in June, the plant did one of its random, unexplained and unexpected changes in energy performance. In Figure 3 you can see the deviation from predicted consumption for intervals when the ECM was active:
The chillers change their behaviour about two-thirds of the way through this history. If we look at the regression model for the ‘ECM on’ condition we can see the unexplained shift very clearly:
So we ended up, in effect, with two tests: one for June and one for July. Looking now at the comparison between ‘on’ and ‘off’ performance in July alone we saw a clear difference:
A similar picture was obtained from observations made in June and the conclusions were that savings of 16.1% and 17.5% were attributable to the ECM in June and July respectively. However, as a further bonus we observed:
The chiller installation’s performance spontaneously deteriorated by 8% at the end of June, echoing behaviour first witnessed in 2019. Identifying the cause will probably save money quite easily; and
Cleaning the condenser coils made no difference to performance. They were probably clean and it was a waste of money, so I suggested not cleaning them until condenser temperatures started to rise.
What lessons would I draw from this episode? That the verifier needs to be vigilant, sceptical, and cautious but flexible. In fact flexibility is needed on all sides, and that is best served by developing trust; trust in this case was built up through openness and continuous communication.
ENERGY surveys and audits – deliberate studies to find energy-saving opportunities – can be done with three levels of depth and thoroughness, can look at three broad aspects of operations, and will generally adopt one of three approaches.
Depth and thoroughness
Let’s take depth and thoroughness first. Level 1 would be an opportunity scan. This typically has a wide scope, and is based on a walk-though inspection. It will use only readily-available data, and provide at most only rough estimates of savings with no implementation cost estimates. It will yield only low-risk recommendations (the “no-brainers”) but should identify items for deeper study.
Level 2 is likely to have a selective scope (based perhaps on the findings from a Level 1 exercise). It is best preceded by a desktop analysis of consumption patterns and relationships, which means first collecting additional data on consumption and the driving factors which influence it. It should yield reasonably accurate assessments of expected savings but probably at best only rough cost estimates. It can therefore provide some firm recommendations relating to ‘safe bets’ and otherwise identify possible candidates for investment.
Level 3 is the investment-grade audit. This may have a narrow scope – perhaps one individual project – and will demand a sketch design and feasibility study, with accurate assessments of expected savings, realistic quotations for implementation, sound risk evaluation and (I would recommend) a measurement and verification plan.
Aspects covered
Next we will look at the three broad aspects of operations that the audit could cover. These are ‘technical’, ‘human factors’, and ‘procedural’.
Technical aspects will encompass a spectrum from less to more intrusive (starting with quality of automatic control and set points through energy losses to component efficiencies). In manufacturing operations the range continues through process layouts, potential for process integration and substitution of alternative processes.
Human-factors aspects meanwhile will cover good housekeeping, compliance with operating instructions, maintenance practices, training needs and enhanced vigilance.
Thirdly, procedural aspects will include the scope for improved operating and maintenace instructions, better plant loading and scheduling, effective monitoring and exception handling, and ensuring design feedback.
Approaches to the audit
The final three dimensions relate to the audit style, which I characterise as checklist-based, product-led, or opportunity-led.
The checklist-based approach suits simple repetitive surveys and less-experienced auditors.
Product-led audits have a narrow focus and exploit the expertise of a trusted technology supplier. Because the chosen focus is often set by advertising or on a flavour-of-the-month basis, the risk is that the wrong focus will be chosen and more valuable opportunities will be missed. Or worse still, the agenda will be captured by snake-oil merchants.
Finally we have the ‘opportunity-led’ style of audit. This is perhaps the ideal, although not always attainable because it needs competent auditors with diverse experience and will include the prior analysis and preliminary survey mentioned earlier.
These ideas, together with other advice on energy auditing, are to be covered in a new optional add-on module for my “Energy efficiency A to Z” course which explais a wide range of technical energy-saving opportunities. Details of all my forthcoming training and conferences on energy saving can be found at https://vesma.com/training.
Note: this guidance was developed because of the Covid-19 pandemic. However, by its nature, the pandemic is an unexpected major non-routine event which will in many cases completely invalidate baselines developed before March 2020. Readers should not expect this advice to remedy the resultant disruption to evaluations. It may prove to be applicable only to ‘retrofit isolation’ assessments and trials using extended sequences of on-off mode changes.
THIS GUIDANCE proposes enhancements to standard measurement and verification protocols to cover the situation where, for reasons beyond the control of the parties involved, a measurement and verification practitioner (MVP) is unable to participate in person.
Firstly the proposal for the energy-saving project should indicate not only the quantity or proportion by which consumption is expected to be reduced, but wherever possible the nature of the expected reduction. For example where data are to be analysed at weekly or monthly intervals, it may be possible to say whether reductions are expected in the fixed or variable components of demand or both; while for data collected at intervals of an hour or less it may be possible to define the expected change in daily demand profile or other parameters related to the pattern of demand.
Setting the expectations more precisely in this manner will help the MVP to detect whether post-implementation results may have been affected by unrelated factors.
Secondly a longer period than usual of pre-implementation data should be collected and analysed. This is necessary in order not only to establish prevailing levels of measurement and modelling uncertainty, but potentially to expose pre-existing irregularities in performance. Such monitoring should employ the same metering devices which will be used for post-implementation assessment.
The causes of historical irregular performance should be traced as they could provide clues about foreseeable non-routine events (NRE) which would then need to be allowed for in post-implementation assessment in case they recur. If NREs are not adequately allowed for, they will at best degrade the analysis and at worst lead to incorrect conclusions.
Thirdly, all parties should remember that as well as foreseeable NREs, there will be unforeseen ones as well. Dealing with is part of standard practice but because the appointed MVP is unable to visit the subject site and interview key personnel, he or she is likely to miss important clues about potential NREs which otherwise would have been evident based on his or her professional experience. It is therefore imperative that a planning teleconference takes place involving local personnel who are thoroughly conversant with the operation and maintenance of the facility. As part of this meeting a knowledgeable client representative should provide the MVP with a clear account of how the facility is used with particular emphasis on non-standard situations such as temporary closures. Pertinent input from client representatives at large would include (to give some examples) control set-points, time schedules, plant sequencing and duty-cycling regimes, occupation patterns, the nature and timing of maintenance interventions and so on. Information about other projects—both active and contemplated—should be disclosed. The MVP has a duty to probe and ask searching questions. It should never be assumed that something is irrelevant to the matter in hand and as a general rule no question asked by the MVP should go unanswered.
It may be helpful to provide the facility of a walk-through video tour for the benefit of the MVP, which can of course be on a separate occasion.
In a recent newsletter I suggested that somebody wishing to monitor the health of their solar PV installations could do so using ‘back-to-back’ comparisons between them. Reader Ben Whittle knows a lot more about these matters and wrote to put me right. His emails are reproduced here:
I would point out that whilst it may possibly be interesting to compare solar installations and account for cloud cover, personally I wouldn’t bother!
variable cloud cover is variable – you can’t control it and generally it is fair to say that annual variation in the UK rarely deviates +/- 5% annually
if you have a monitoring system, it will be capable of telling you when there is a fault immediately by email rather than waiting to do analysis
In the case of inverter manufacturers’ own monitoring systems they will directly report faults immediately , usually based on
an actual fault code being generated by the inverter – typically either being switched off and failing to report at all, string insulation resistance faults or other major failures
output not matching other inverters in the same installation or sometimes against a base case / prediction based on yield expected due to weather forecasts
possibly against a self defined target, and a failure to meet it
Third-party monitoring manufacturers will typically do the same as inverter manufacturer monitoring (with the exception of not reporting actual fault codes), but they have the advantage of being able to report on installations with mixed inverter manufacturers being used (possibly new + historic installations in one location or a portfolio of installations from different installers)
One classic mistake made with solar monitoring information is having no clear idea of what and how you are going to deal with all the information! It is time consuming to do and takes a bit of experience to make sense of it all.
So I asked Ben if there was a cost attached and this was his reply:
Most inverter manufacturers provide a solution. Three of the biggest brands (SMA, Fronius, SolarEdge) all have very competent systems, which are hosted for free, but you can get additional services by paying extra.
A typical domestic setup (which could cover any installation to any size in theory) would have basic info on annual, monthly and daily yield, and may also display self consumption rates assuming you have bought the requisite sub meter. Other info can include energy sent to a battery if you have one. It would also notify you if you lost grid connection, or communication faults. Communication is typically managed over wifi for domestic set ups and ethernet in commercial set ups. Remote solar farms do all this over 3g or 4g if there is no nearby telephone infrastructure.
Where you would pay money for a service is for an enterprise solution: this would allow you to also compare multiple installations and give you more detailed performance info, possibly also automating equipment replacement or engineer visits if malfunctions were being detected. (You would only get this from the major manufacturers with a dedicated team in this country, or an O&M service provider who was being paid to keep an eye on performance).
Third party systems typically only work using generation meter and export meter info, but a surprising amount of knowledge can be gleaned from this – you are after all only trying to find anomalies and once you have defined the expected performance initially this is quite straightforward. The advantage of this is that if you are managing lots of different installations with different inverters then you can pull the data all into one database. Big O&M companies may insist on this being added where a service level is being defined – such as 98% availability or emergencies responded to in under 24 hours. The service will also include additional data points such as pryanometer info and other weather data, depending on the scale of the installation.
The companies who operate big solar farms are often hedge funds and they don’t like leaving systems down and not running for any length of time given the income from feed in tariffs. Though they quite often don’t manage the farms as well as they could do…
Ben Whittle (07977 218473, ) is with the Welsh Government Energy Service